Udo Kuckartz, Stefan Rädiker - Qualitative Content Analysis - Methods, Practice and Software-SAGE Publications (2023)
Udo Kuckartz, Stefan Rädiker - Qualitative Content Analysis - Methods, Practice and Software-SAGE Publications (2023)
CONTENT ANALYSIS
2nd
Edition
UDO KUCKARTZ
STEFAN RÄDIKER
ISBN 978-1-5296-0914-1
ISBN 978-1-5296-0913-4 (pbk)
At SAGE we take sustainability seriously. Most of our products are printed in the UK using responsibly
sourced papers and boards. When we print overseas we ensure sustainable papers are used as measured
by the PREPS grading system. We undertake an annual audit to monitor our sustainability.
List of Figures ix
List of Tables xi
Preface xiii
About the Authors xvii
2 On Categories 33
3 Developing Categories 49
Figure 5.1: Excerpt from the interview guide of the example study 101
Figure 5.2: The seven phases of structuring QCA 102
Figure 5.3: List of main thematic categories 104
Figure 5.4: T
hematic cases-by-categories matrix as a starting point (top)
for thematic case summaries (bottom) 112
Figure 5.5: Coded text passages as a starting point for thematic summaries 113
Figure 5.6: Analysis options after completion of coding in structuring QCA 115
Figure 8.1: T
ranscription rules for computer-assisted analysis and notes on
their implementation in automatic transcription software 163
Figure 8.2: Excerpt from a transcript (interview with participant R7) 164
Figure 8.3: Interview excerpt with open codes displayed on the left 172
Figure 8.4: Excerpt from interview R29, paragraphs 33–37 172
Figure 8.5: Paraphrases next to the text 173
Figure 8.6: Summarizing coded segments using the Summary Grid 174
Figure 8.7: Visualization of the progress of a group discussion 186
Figure 8.8: Visual representation of interviews and assigned categories 186
Figure 8.9: Keyword-in-context for the word ‘family’ (excerpt) 188
Figure 8.10: Word tree with focus on the word ‘family’ (detail) 189
Table 5.1: D
efinition of the sub-categories within the main category
‘biggest problems in the world’ 108
Table 5.2: Tabular case overview (excerpt) 117
Table 9.1: Quality standards within quantitative and qualitative research 194
Table 9.2: Coding table for ten coding units, two coders, and one category 198
Table 9.3: Table of matches for one category and two coders (general form) 198
Table 9.4: Table of matches for one category and two coders 199
Table 9.5: Table for two coders and four categories 200
Table 9.6: Comparison of coding for two coders and multiple categories 202
Table 9.7: C
omparison of frequency of coding for two coders and
several categories 203
Table 9.8: Match table for two coders 204
Acknowledgements
We would like to thank all those who have contributed to this book since the publica-
tion of the first German edition and who have given us many further ideas and valuable
feedback, especially the many students and young academics who have asked challeng-
ing questions or criticized our approach in seminars, workshops, and summer schools.
We are especially grateful to Graham Hughes for his invaluable support in the English
translation of this book. It is always a great joy when a book is finally done after many
months of work. However, writing a book is usually a never-ending story. When brows-
ing through a freshly printed copy for the first time, ideas for improvement arise and
sometimes a few little mistakes are discovered right away that have crept in despite all
the effort. We would like to encourage all readers of this book to give us feedback: sim-
ply send us your comments and suggestions by e-mail to [email protected] or
[email protected].
Online Resources
We have set up a website at www.qca-method.net that includes many useful resources
for students and instructors, for instance PowerPoint slides, lecture notes, discussion
questions, suggested exercises, and selected illustrations from this book for download.
evaluation. His aim is to teach the application of research methods in a structured and
easy-to-understand manner. In addition to several book chapters, he has co-authored
and co-edited several books on qualitative and quantitative research methods, among
others Qualitative Evaluation, The Practise of Qualitative Data Analysis: Research Examples
using MAXQDA, Analyzing Qualitative Data with MAXQDA: Text, Audio, and Video and
Focused Analysis of Qualitative Interviews with MAXQDA: Step by Step.
Chapter objectives
In this chapter you will learn about:
analysis’ which is immediately apparent when the three words ‘qualitative’, ‘data’, and
‘analysis’ are bracketed together in different ways. While ‘(qualitative data) analysis’
refers to the analysis of qualitative data in the above sense of texts, images, films, etc.,
‘qualitative (data analysis)’ means the qualitative analysis of all kinds of data, that is,
both qualitative and quantitative data. Differentiating between data and analysis results
in a four-cell table1 as presented in Table 1.1.
Table 1.1 Qualitative and quantitative data and analysis (according to Bernard & Ryan, 2010, p. 4)
The upper left cell A and the lower right cell D appear well known to us. Cell A con-
tains the qualitative analysis of qualitative data, for instance in the form of hermeneutic
analyses, grounded theory, or other qualitative analysis methods. Cell D, the quantitative
analysis of quantitative data, is also familiar to us. This involves using statistical methods,
that is, the typical process of analysing numerical data. However, the table also includes
two unexpected combinations, namely the qualitative analysis of quantitative data (cell
B) and the quantitative analysis of qualitative data (cell C). The latter may include, for
example, the analysis of word frequencies and word combinations. The qualitative
analysis of quantitative data (cell B), which involves interpreting quantitative data,
begins when the statistical procedures have been done and the results are available in
the form of tables, coefficients, parameters, and estimates. At this point it is time to
identify and interpret the meaning of the results and to work out their substance.
Without this qualitative analysis work, the mere numbers remain sterile and literally
meaningless. As Marshall and Rossman emphasized, the interpretive act is inevitable:
1
The table is based on the earlier differentiation by Alan Bryman (1988) of qualitative
and quantitative research rather than qualitative and quantitative data. Bryman classified
cells B and C of the table as ‘incongruent’.
Bernard and Ryan’s differentiation makes it clear that the type of data does not necessarily
determine the type of analysis. If one moves away from such a strict connection between data
type and analysis type, it is clear that both a quantitative analysis of qualitative data and a
qualitative analysis of quantitative data are possible. Thus, there is no reason to assume a deep
divide between the qualitative and quantitative perspectives. In everyday life, as in science, we
humans have a natural tendency to combine methods. We humans always try to keep both
perspectives – the qualitative and the quantitative aspects of social phenomena – in mind.
In the latest edition of their handbook on qualitative research, Denzin and Lincoln
emphasize the diversity of qualitative research, which shows how impossible it is to
provide a ‘one-size-fits-all’ definition:
2
At least, this is the impression one gets when reading Denzin and Lincoln (2018) or the
abstracts of the Qualitative Inquiry conferences (www.icqi.org).
the diversity of approaches to qualitative research. The result was a tableau of almost
50 different qualitative approaches, trends, and forms of analysis, ranging from ‘action
research’ to ‘transformative research’ (Tesch, 1990, pp. 58–59). Tesch arranged the var-
ious approaches in a cognitive map and differentiated them according to whether the
research interests were focusing on the characteristics of language, the discovery of reg-
ularities, understanding the meaning of the text or the act, or reflection.
It seems as if almost every author of a textbook on qualitative methods feels committed
to creating a new systematization of qualitative approaches. The results of such systema-
tizations differ greatly. For example, Creswell and Poth (2018) distinguish five (main)
approaches of qualitative research: ‘narrative research’, ‘phenomenology’, ‘grounded the-
ory research’, ‘ethnography’, and ‘case study’. In contrast to Tesch’s differentiation based
on research interests, Creswell and Poth focus on epistemological and pragmatic aspects.
This is not the place for a synopsis of this multitude of systematizations; we merely
point out the existence of a great variety of qualitative approaches that do not share a
uniform theoretical and methodological understanding (Flick, 2007, pp. 29–30).
Accordingly, the definitions of ‘qualitative research’ vary greatly. Some elements,
including case orientation, authenticity, openness, and integrity, can be found in
almost every definition. It will suffice here to refer to the 12 characteristics of qualitative
research practice listed by Flick et al. (2017, p. 24):
What shines through in Oswald’s position, namely that qualitative and quantitative
methods are not mutually exclusive, is the focus of the discourse on mixed methods.
Mixed methods approaches are – as the leading actors argue – a new contemporary
understanding of methods that tries to overcome the old duality of approaches in a
new, third paradigm. Scholars such as Bazeley (2018), Creswell and Plano Clark (2018),
Mertens (2018), Morgan (2014), and Tashakkori et al. (2021) have elaborated mixed
methods approaches in detail and developed a variety of precise design proposals for
mixed methods research. In terms of research practice, the proposals of these authors are
extremely interesting and relevant for research projects in many scientific disciplines.
Methodologically, Udo Kelle’s work to integrate methods should be taken into account
in this context (Kelle, 2008). While the mixed methods movement is committed to
pragmatism (Creswell & Plano Clark, 2011, pp. 22–36), Kelle’s (2008) approach is episte-
mological, beginning with the controversy regarding the role of explanation and under-
standing that shaped the humanities and natural sciences for more than 100 years. His
concept of the integration of methods is methodological and he attempts to substan-
tiate the combination of methods at a much deeper level. Kelle goes back to the dawn
of empirical social research and the qualitative–quantitative controversy, and asks how
it is possible to develop empirically-based theories in the social sciences and arrive at a
concept of ‘causal explanation’, which, in principle, was already present in Max Weber’s
research (Kuckartz, 2009).
In an online German doctoral forum, for example, a graduate student posted the
following plea for help:
Hello,
Hmm. Now I am rummaging through a lot of literature, mostly from the social
sciences. But I simply cannot find anything tangible for analysing qualitative
data. This is all very vague. And I would really like to report some results at the
end. Feeling a little hopeless at the moment. Can anyone here give me any tips?
Regards,
Dana
Dana is right: a tangible and concrete method for analysing qualitative data is not easy
to find. And that is where this book comes in: our aim is to show systematically and
methodically ways in which qualitative data can be analysed. Collecting qualitative data
is not only interesting and exciting but also usually feasible without major methodolog-
ical problems. The difficulties with which researchers are faced in the early stages of a
project are more related to field access or one’s own behaviour in the field, rather than
the methods employed to collect in the narrower sense. But what comes after you have
collected the information, when the interviews or videos have been recorded and the
field notes are written?
Students are not the only ones who feel unsure at this point in the research process,
and many avoid the risks associated with qualitative research, because the analysis pro-
cess and its individual steps are not described precisely and in enough detail in the
literature and are therefore difficult to carry out. Even in the reports of large-scale funded
research projects, there are often only very imprecise descriptions of the approach to data
analysis. Researchers often use empty phrases or merely describe that they ‘based their
analysis on the grounded theory’, ‘interpreted according to Silverman’, ‘on the basis of
qualitative content analysis’, or by ‘combining and abbreviating different methods’. A
precise, well-understandable representation of the procedure is often omitted.
On the other hand, the mentality of ‘anything goes’3 can often be found in the dis-
course on qualitative data analysis methods. Researchers who read qualitative methods
3
The slogan ‘anything goes’ of the American philosopher of science Paul Feyerabend was
not meant as a licence for researchers to do anything they wanted methodically
speaking, but as an invitation to use creative methods in their research.
texts and come to such a conclusion believe that they can more or less do what they
want, make glorious interpretations, and let their own imaginations and associations
have free rein, without the danger of strict methodologists rejecting them and/or put-
ting them in their place. They can even call on the constructivist and postmodern
positions encountered in the discussion of the quality standards for qualitative research,
which emphasize that the social world itself is constructed cognitively and that multiple
worlds and world-views exist side by side; thus, the question of universal and objective
quality standards can be regarded as obsolete. Such positions are not shared in this
book. For us, Seale’s position of a ‘subtle realism’ (Seale, 1999a) is convincing: in the
discourse on the quality of qualitative research, Seale pleaded pragmatically (building
on Hammersley, 1992) for a compromise between the two extremes, namely between
the adherence to the rigid rules of classical research concepts (objectivity, reliability,
validity) on the one hand and the rejection of general criteria and standards on the
other. Promoting the formulation of appropriate quality standards and precise descrip-
tions and documentation of analytical procedures (see Chapter 9) would undoubtedly
increase credibility and reputation when addressing a ‘sceptical audience’ (Seale, 1999a,
p. 467) as well as research-funding institutions.
4
It is a language spoken in the East African country of Rwanda and the eastern Congo.
example, what the euro, the different countries in the EU, and the different financial
policies are in order to understand the article. Finally, in order to really understand it,
you have to know the history of the euro and be familiar with the aims of having a
single currency in the EU.
The more we know, the better we are able to recognize that a text has different levels
of meaning. For example, only with previous knowledge on the subject could you rec-
ognize that the politician quoted in a newspaper article who used to be a strict opponent
of financial support for Greece has now given surprisingly balanced and convincing
arguments in favour of such support. Moreover, if you know that that same politician
is an active member of the state government, you can assume that that governmental
body may be changing its stance on the issue as well.
It is impossible to gain an inductive understanding of a text by itself. Middle Age
biblical illustrations serve as a good example of this: the more you know about the
iconography of the time and the better your knowledge of Christian symbolism, the
better you will understand a given illustration. This sort of understanding cannot
be deduced from the illustration alone, as Christian symbolism goes beyond the
illustration – and the Bible cannot be construed inductively based on illustrations
of different biblical scenes.
Important points of orientation for the analysis of qualitative data are general con-
siderations about understanding, specifically understanding and interpreting texts. In
the German-speaking world, this is often associated with hermeneutics. But what exactly
is meant by hermeneutics? What does this term mean, which hardly plays any role in
the Anglo-Saxon social science methodological literature?
The term ‘hermeneutics’ is derived from the Greek word ἑρμηνεύειν, which means to
explain, interpret, or translate. Hermeneutics, then, is the art of interpretation, the
techniques involved in understanding written texts. As a theory of understanding,
hermeneutics has a long history that extends as far back as the medieval interpretations
of the Bible or even to Plato. Within the context of scientific thought, hermeneutics
appeared in the late nineteenth century as leading philosophers, including Schleiermacher
and Dilthey, proposed it as the scientific approach of the humanities in contrast to the
explanatory methods of the natural sciences. Cultural products such as texts, illustra-
tions, pieces of music, or even historical events were to be developed and understood
within context. Dilthey wrote that we explain nature, but in the human sciences we
have to establish a different methodological foundation based on understanding and
interpretation (Verstehen). Dilthey’s famous sentence ‘We explain nature, we understand
psychic life’ is programmatic (Dilthey, 1894/1977, p. 27).
The contrast between explaining and understanding has been discussed a great deal
in the literature on the philosophy of science and we will not address it any further here.
If you are looking for an instructive text on the topic, see Kelle (2008, pp. 159–164), who
tries to overcome the opposition of explaining versus understanding with a new
approach. Kelle relies on the concept of multiple causality developed by the Australian
philosopher John Mackie.
Over time, hermeneutics has evolved – from Schleiermacher and Dilthey to the
modern-day approaches of Gadamer, Klafki, Mollenhauer, and others,5 there is no single,
uniform hermeneutical approach today. Some time ago, Anglo-American philosophers
also became aware of hermeneutics through the work of Richard Rorty (1979). For the
purposes of this book, we are interested less in the historical, theoretical, and philo-
sophical aspects of hermeneutics and more in the guidelines hermeneutics offers for the
analysis and interpretation of data collected in qualitative research projects. How do we
take a hermeneutical approach to analysing the content of texts? Klafki presented a
comprehensive example based on an interpretation of a Humboldt text about how to
construct the Lithuanian city school system (Klafki, 1971/2001). In his text, Klafki
formulated 11 methodological insights for his hermeneutical approach, which still apply
today. Four of the main points are important within the context of QCA.6
First, pay attention to the conditions under which the text was created. Bear in mind
the conditions under which the text to be analysed (e.g., an open interview) was cre-
ated. Who is communicating with whom, under what circumstances? How much and
what kind of interaction did the researcher have with the field prior to the interview?
How would you characterize the interaction between interviewers and interviewees?
What information have the research participants received about the project in advance?
What are the mutual expectations? What role does social desirability in the interaction
possibly play?
Second, the hermeneutic circle. The central principle in the hermeneutic approach
is that a text can only be understood as the sum of its parts and the individual
parts can only be understood if you understand the whole text. One approaches
the text with a pre-understanding, with assumptions about the meaning of the
text, reads it in its entirety, works through the text, which leads to a further develop-
ment of the original pre-knowledge – always provided, of course, that one shows
openness in working through the text and is prepared to change previously existing
judgements.
Any attempt to understand a text presupposes some prior understanding on the
part of the interpreters. Klafki noted that reading through the text and/or parts of
the text multiple times results in a circular process (Klafki, 1971/2001, p. 145); how-
ever, it would seem that a spiral serves as a more suitable illustration since you do
not circle back to your starting point. Instead, you develop a progressive under-
standing of the text. The hermeneutic circle or spiral is often visualized as shown
in Figure 1.1.
5
Gadamer elaborated a concept of philosophical hermeneutics; in his book Truth and
Method (2013) he dealt with the nature of human understanding.
6
In this section, we draw on central statements on hermeneutics in Vogt (2016).
Prior Understanding
P2 P1 U1 U2
understanding of the text
Third, the hermeneutic difference. The notion of the hermeneutic difference points
to the central problem of all verbal communication, namely that we can only under-
stand texts and communication in general – or think we understand them – through an
interpretive process. The hermeneutic difference can vary greatly in degree. It is very
high when, for example, we visit a foreign country and cannot understand the language
that is spoken, even higher when – as in Chinese – the character system is foreign to us,
and we cannot even look up the unknown words in a dictionary.7 In everyday commu-
nication, the hermeneutic difference seems small or even obsolete to us. According to
Schleiermacher, no hermeneutics is necessary to talk about the weather or when we
order ‘Five rolls, please’ at the bakery. As Gadamer noted, hermeneutics takes place in
the grey area between foreign and familiar: ‘Hermeneutics is situated in this place in
between’ (Gadamer, 1960, p. 279).
Four: Accuracy and suitability. Hermeneutic procedures attempt to understand cul-
tural products such as texts, images, and art. As Mollenhauer and Uhlendorff (1992)
emphasize, they attempt to understand accurately. However, no methodology can guar-
antee accuracy. In hermeneutics, it all depends on the person trying to understand or
interpret something, and each person always has some sort of preconception about the
object or subject at hand. Gadamer stressed that these are preconceptions or assump-
tions. Thus, a hermeneutic interpretation that fulfils the criteria for intersubjective
agreement cannot be postulated per se. There is no right or wrong interpretation, only
a more or less suitable interpretation.
7
Generally, we can distinguish between three forms of hermeneutic difference: linguistic,
historical, and rhetorical. In the example above, it is a linguistic difference. Historical
difference can manifest itself as a factual or linguistic difference, such as in the form of
outdated terms or sayings, or unknown persons, facts, and situations.
1 Reflect on your own preconceptions and any assumptions you may have
regarding the research question.
2 Work through the text as a whole, setting any unclear passages of the text aside
until you gain a better understanding of the entire text which may shed light on
the unclear passages.
3 Make yourself aware of the hermeneutic difference by asking yourself, ‘Does the
text contain a different language or culture with which I am unfamiliar?’ Try to
reduce these differences, such as by learning the new language or finding an
interpreter.8
4 During your first reading of the text, pay attention to the topics or themes
appearing in the text which are important to your research.
5 Differentiate between a logic of application (i.e., the identification of existing
themes and categories in the text, as when the text is indexed) and a logic of
discovery (i.e., the identification of important new, perhaps even unexpected
things, in the text).
8
This is true in cross-cultural research, but it can also be useful for research conducted
in a familiar environment. Sprenger (1989) tells of a social science project about the use
of technology in critical care and how medical experts were invited to help the research
team interpret the phenomena they observed, which made a scientific analysis possible.
They are not another way to answer the same question. Instead, they constitute
a relatively new way to answer a different type of question, one characterised
by a unique approach with a different set of underlying assumptions reflecting
a different worldview of how individuals and group behaviour can best be
studied. (Miller & Salkind, 2002, p. 143)
Diekmann makes a somewhat more differentiated distinction between the forms of empir-
ical studies. He distinguishes four types of studies (2007, pp. 33–40): explorative, descrip-
tive, and hypothesis-testing studies and, as a fourth type, evaluation studies. Both qualitative
and quantitative methods can be used in all four types of study, and it is also possible to
combine both methods within one type of study. According to Diekmann, the proportion
of qualitative methods is different for the different types of studies. While mostly qualita-
tive methods can be found in exploratory studies, descriptive studies, which will give the
most generalized overview possible, rely on more quantitatively oriented survey research.
The starting points for all of the above forms of research are the research questions.
Without such questions, research is difficult to imagine. Because no matter whether you
are planning a bachelor’s, master’s, or doctoral thesis or you are writing a research pro-
posal to receive third-party funding, the first step is always to face the challenge of
drawing up an exposé, a research plan, or research proposal, in which the presentation
and discussion of the research question plays a central role.
When formulating research questions, you should always reflect on the theoretical back-
ground and your own prior knowledge, that is, ask yourself: How much have I thought
about this field of research. What research already exists? Which theories seem to have
explanatory power regarding my research questions? What prejudices do I have myself
and what prejudices are common among the scientific community of which I am a part?9
9
Those looking for further suggestions on how to formulate research questions will find
them, among others, in Creswell and Creswell Báez (2021, pp. 95–104), Creswell and
Poth (2018, pp. 127–146), Flick (2018a, pp. 83–95), and O’Leary (2018).
To ask such questions is not in conflict with the idea of openness that is character-
istic of qualitative research. The common assumption that researchers can be a ‘tabula
rasa’ or a ‘blank slate’, able to devote themselves to a research subject entirely without
prior knowledge is an illusion (Kelle, 2007). Prior knowledge is always a factor, as the
researcher’s brain is never ‘empty’. Even if, after well-founded consideration, you
choose not to refer to existing research results because you would like to approach your
research question and approach the field ‘without prejudice’, you should reflect on
your reasons for doing so and record them on paper. A mere reference to scholars who
recommend such a theory-free and unprejudiced approach is not sufficient to justify it;
instead it requires reflection regarding exactly why such a theory-abstinent approach
to answering your research question is appropriate and why this promises better results.
It is not uncommon to come across statements referring to grounded theory, according
to which reading books on the topic of the research is said to be counter-productive in
terms of research methodology. This is grotesque nonsense that is at best suitable for
discrediting qualitative approaches in the scientific community and the wider public.
In grounded theory itself, this misunderstanding found in the reception of the early
grounded theory texts (Glaser & Strauss, 1998) has long been corrected (Corbin inter-
viewed by Cisneros-Puebla, 2004; Kelle, 2007; Strauss interviewed by Legewie &
Schervier-Legewie, 2004).
Of course, there are situations in social research in which it is advantageous to gain
experience in the field first. For instance, anyone who wants to observe and experience
how homeless people live should not simply plan to sit in the library reading the socio-
logical and psychological literature on homeless people. However, it would make sense
to consider the state of research either following the observation and in the course of the
data analysis, or at the latest when discussing the results. On the other hand, it is hard
to imagine that anyone who wants to analytically explore the causes of right-wing think-
ing in adolescents would consistently ignore all of the research literature that already
addresses that very problem. In this book, the position is taken that it is wise and neces-
sary to start with the existing research when exploring social phenomena. We agree with
Hopf and Schmidt who encouraged researchers to delve into the current state of research
on the chosen topic:
validity have been discussed intensively. Three principal positions are taken regarding
the acceptance and transferability of existing quality standards for quantitative research:
• Universality. The same standards are valid for qualitative research as for
quantitative research.
• Specificity. Specific standards that are appropriate must be formulated for
qualitative research.
• Rejection. Quality standards are generally rejected for qualitative research.
Flick (2018a, pp. 559–560) adds a fourth position, namely, that researchers should be
able to answer the question of quality beyond the formulation of standards, such as
in the form of total quality management, which takes the entire research process into
account. For the general discourse on the quality standards for qualitative research, it
will suffice here to refer to relevant contributions (Flick, 2020; Seale, 1999a; Steinke,
1999). In this book, the topic will be considered through a focus on the method of
qualitative content analysis, and the second of the above positions will be used as the
basis, namely that specific, appropriate standards for qualitative research must be formu-
lated and not simply carried over from quantitative research. Inspired by psychological
test theory, standards for objectivity, reliability, and validity have been established in
quantitative research, which can be found in almost every textbook of social research
methods. These quality standards are based on the scientific logic of measurement and
are more oriented towards measurable variables (e.g., reliability coefficient). Standards
for the quality of qualitative research, however, cannot be based on calculations and
measures, as the data for such a calculation is missing. Thus, following Flick (2018a),
the standards themselves must be more process-oriented.
In recent years, increased efforts have been made to canonize qualitative research
procedures and to discuss aspects of quality (Flick, 2018a, pp. 539–568). In particular,
the work of Clive Seale has been given a lot of attention. Seale and Silverman (1997, p. 16)
pleaded, as shown above, in favour of ensuring rigour in qualitative social research and
the construction of quality criteria. Does this mean that we have to take over the logic
behind the quality standards of quantitative research and apply fixed technical evalua-
tion instruments? Seale’s position of ‘subtle realism’ is a middle way, beyond loose
acceptance or rejection of the classical quality standards. The standards within quantitative
research cannot be carried over directly to qualitative research.
Qualitative research is carried out in natural settings and differs from the hypo-
thetico-deductive research model. There, the focus is on testing hypotheses and the goal
is to find correlations and create causal models that can be generalized. Qualitative
research can generalize, too, but this is not its main purpose. In particular, the broad
generalization that is inherent in the research logic of the hypothetico-deductive model
is a foreign concept in qualitative analysis (Seale, 1999a, p. 107). Ultimately, the goal of
the hypothetico-deductive model is to discover patterns and even laws with universal
and long-term validity, while in qualitative research, in particular in the theory-building
grounded theory approach, the goal is to establish middle-range theories.
What, specifically, are the reasons for proceeding with methodological rigour when
analysing qualitative data? Five aspects are important arguments for systematic kinds of
analysis and qualitative text analysis in particular:
The application of methodological rigour also deals with the problem of quantification
in qualitative research:
Yet, as I showed in the last chapter, numbers have a place within qualitative
research, assisting, for example, in sensitive attempts to learn lessons in one
place that have relevance for actions and understanding in another place.
There is a variety of other uses of numbers which can enhance the quality of
qualitative research … (Seale, 1999a, p. 120)
As a result of his very instructive overview of the benefits and use of numbers in qual-
itative research (Seale, 1999a, pp. 119–139), Seale formulated the principle of ‘count-
ing the countable’. Numbers can assume different functions: they can represent not
only simple frequencies or percentages, but also be used for more complex statisti-
cal calculations, such as crosstabs with the chi-square test or cluster analysis. They
can clarify arguments and support theories and generalizations. Seale’s emphasis on
‘avoiding anecdotalism’ expresses the importance of using numbers quite concisely
(Seale, 1999a, p. 138).
Weber’s suggestion included four aspects that were quite characteristic of the subsequent
development of content analysis:
What makes classical content analysis so interesting for the development of methods for
systematic qualitative text analysis is that it is based on nearly 100 years of experience
in systematically analysing texts, even a large quantity of texts. This means that it has
already encountered (and often solved) a variety of problems that arise when analysing
written texts or verbal data, which in fact are qualitative data.
Scholars like Krippendorff and Merten note that the history of content analysis
began a long time ago. Merten sees precursors of content analysis in the exegesis of the
Bible or in Sigmund Freud’s interpretation of dreams. Within this context, Merten
mentions an ‘intuitive phase’, which extends to approximately 1900 (Merten,
1995, pp. 35–36). The actual beginning of scientific content analysis is dated around
the beginning of the twentieth century and marked by Max Weber’s speech at the first
German Sociological Association Conference, and his proposal for an ‘inquiry of the
newspapers’, as noted above. Numerous studies and analyses were completed in the
field of communications in this ‘descriptive phase’. The golden age of content analysis
came with the invention of the radio and particularly with the analysis of the impact
of reporting on the war during the 1940s. Famous projects, such as the ‘World Attention
Survey’ in 1941 and Lasswell’s study of war reports and propaganda (for the Experimental
Division for the Study of Wartime Communication, sponsored by the US government
and the Hoover Institute), make it evident that content analysis in the field of commu-
nications was also politically important at the time. Under the leadership of Lazarsfeld
and in occasional cooperation with Adorno, the Rockefeller Foundation’s outstanding
‘Radio Project’ researched the effects of mass communication (‘propaganda analysis’).
The term ‘content analysis’ was first used in 1940 and many other terms that are
central to content analysis, such as ‘sampling unit’, ‘category’, and ‘intercoder reliabil-
ity’, stem from that time and were coined by leading content analysts like Lasswell,
Berelson, and Lazarsfeld. Methodically, content analysis made considerable progress:
Berelson wrote the first dissertation using methods of content analysis in 1941 and later
co-authored the textbook The Analysis of Communication Content with Lazarsfeld
(Berelson & Lazarsfeld, 1948). In addition, numerous publications and conferences
made it possible for researchers to exchange their ideas and methodologies (Früh, 2017,
pp. 11–15).
Since the end of the 1940s, content analysis has taken on more of a quantifying and
statistical character. This must be viewed within the context of a general shift in the
social sciences towards behaviourism after the Second World War and into the 1950s
and the early 1960s. Empirical research focused on testing hypotheses and theories.
Qualitative research was considered unscientific and more and more qualitative ele-
ments disappeared from content analysis, which was then limited to the quantitative
analysis of the manifest content of communication. Thus, Berelson defined content
analysis as follows:
Critique of such a methodically narrow form of content analysis soon followed. Kra-
cauer, for example, criticized Berelson’s content analysis of being too superficial and not
grasping the more subtle meanings. It was Kracauer who first advocated a ‘qualitative
content analysis’ (Kracauer, 1952). Such a qualitative type of content analysis should
also address latent meaning, not in the sense of objective meaning, of probable and
improbable readings, but as latent meaning that can be intersubjectively agreed upon.
This raises the general question of understanding the meaning of texts, for which a
consideration of hermeneutics as the classical theory of interpretation is recommended
(cf. Section 1.4).
A ‘qualitative content analysis’, such as Siegfried Kracauer’s counter-position to the
mainstream content analysis of that time, was a type of content analysis that wanted to
put an end to the restriction to manifest content imposed by the prevailing behaviourist
paradigm. Kracauer argued for a content analysis that also takes into account latent
aspects of the meaning of texts, or of communication content in general (Kracauer,
1952). The QCA that we conceptualize in this book refers on the one hand to historical
social science pioneers as Kracauer, who did not want to limit the method to the manifest
content and its quantification, and on the other hand to hermeneutic traditions, from
which one can learn a lot about the basic principles of understanding text and meaning.
What are the stages that mark the way to the method of QCA presented in this book?
Kracauer had already conceived the method of QCA not as an antithesis to classical
content analysis, but as a necessary extension of content analysis, which had become
increasingly narrow in the course of time. Leading content analysts of that time had
argued that different kinds of texts are situated on a continuum. Statements that do not
require additional interpretation, such as facts or alleged facts, are situated at one end
of the continuum while texts that require interpretation are situated at the other end.
For example, the news of a train accident would be at the factual end of the continuum
while an example of modern poetry would be at the other end.
manifest latent
News Item Poem
“Train accident” modern Iyric
Kracauer argued, however, that in social science analyses events such as train acci-
dents that do not require additional interpretation are very rare. In such cases, counting
and statistical analysis are indeed appropriate and useful. But even beyond the interpre-
tation of modern poetry text analysis is not possible without subjective understanding
and interpretation of text. The crucial point is that quantitative approaches are not as
accurate as interpretive approaches when it comes to understanding communication.
This can be seen, for example, if the attempt is made to rate a complex communication
on a scale with only three points from ‘very favourable’ to ‘very unfavourable’ (Kracauer,
1952, p. 631).
Kracauer praised QCA as a necessary extension of and supplement to mainstream
content analysis, which was becoming more and more quantitative this time. He con-
cluded that a new kind, a qualitative content analysis had to be established. In the
following decades, QCA played only a marginal role in the methods literature. If it was
dealt with at all, it was done in a rather dismissive manner, as Devi Prasad (2019)
describes. However, many qualitative researchers have used content analysis techniques
in their research practice and have thus put Kracauer’s advocacy of QCA into practice.
For decades, methodological development took place in research practice rather than in
the methods literature. The marginalized role of QCA then changed with the emerging
discussion of qualitative research in the 1970s and 1980s. Texts dealing with QCA now
appeared, especially in German-speaking countries (e.g., Ritsert, 1972). Of particular
importance in this context is a book on QCA by Philipp Mayring (1983). This book was
widely read, so that the term ‘qualitative content analysis’ was often associated with
a different definition is needed, one that clearly differs from the definitions of quan-
titative content analysis that can be found in the literature (Früh, 2017; Krippendorff,
2018; Rössler, 2017).
Stamann et al. (2016) have compiled numerous definitions and attempted defini-
tions of QCA. They identify the core of QCA as ‘the systematic analysis of the meaning
of material in need of interpretation by assigning it to the categories of a category sys-
tem’ (2016, para. 5) and formulate the following definition:
ment; secondly, that the analysis is primarily qualitative but can also integrate
quantitative-statistical evaluations; it can be both
primarily qualitative; and thirdly,
category-oriented and case-oriented.
that QCA not only is category-
oriented but also can have a case
orientation. A fourth difference
seems less relevant to us, but should be mentioned: we do not work with an extended
concept of text that also includes videos, films and images, but differentiate between
these types of data.
This definition also differs significantly in some points from the definition formu-
lated by Margrit Schreier:
We see another significant difference in the fact that – as in the definition by Stamann
et al. – Schreier’s definition does not include a case-oriented perspective, which, unlike
in quantitative research, plays a very important role in qualitative social research.
In Mayring (2015), who now prefers the term ‘qualitatively oriented category-guided
text analysis’ instead of QCA (Mayring, 2020, p. 500), there is no definition in the actual
sense, but a list of 15 principles for the development of QCA (2015, p. 49):
10 Possibility of re-interpretation
Such a list of ‘principles’ far exceeds the usual framework of definitions, whose function
would be the concise determination of the essence of something, here QCA. Mayring’s
collection of 15 principles, on the other hand, comes from a tour d’horizon through
different traditions and research approaches, which of course immediately raises the
question of their compatibility (Stamann et al., 2016, para. 6). This will be discussed in
more detail below in Section 1.9.
However, the relevance of a definition should not be set too high. Here it is worth
recalling Karl Popper’s rejection of the traditional view that one must first define terms,
that is, reach agreement on the vocabulary to be used, before beginning a discussion. ‘It
is not by definition that the use of a term is established, but the use of the term estab-
lishes what is called its “definition” or its “meaning”. In other words, there are only
definitions of use’ (Popper, 1979/2010, p. 447).
In this respect, it is more informative for the assessment of the variant of QCA pro-
posed by Mayring to trace his practical use of the method instead of the definition. This
can be understood both from the description of a practical example that the author
included in the 12th edition of his German book (Mayring, 2015, pp. 88–89), and from
the defined flow logic of the various analysis steps in the QCAmap software designed
by Fenzl and Mayring (2017).
In addition to the analysis of these data, which are prototypical of qualitative research,
QCA can also be successfully used for the analysis of other data, such as:
• Podcasts
• Internet data (e.g., blog posts, company websites)
• Scientific publications
• Textbooks.
Interviews and focus groups are usually analysed in written form and not as audio
recordings, that is to say, in this case it is first necessary to transcribe the audio record-
ing. With regard to the types of data, no order of appropriateness can be stated, for
example, in the sense that QCA would be better suited for the analysis of interviews than
for the analysis of documents or social media data. In social research, QCA is very often
used for the analysis of interviews and focus groups, but this does not mean that it is
less suitable for documents or social media data. Whenever qualitative material needs to
be analysed systematically and with the help of categories, QCA is an excellent method.
If you look at the latest volumes of scientific journals, you will see that social media
data are playing an increasingly important role in research. This has consequences for
QCA, because in the analysis of these types of data, considerably more data are to be
expected than is the case in a study with open interviews or focus groups. Given the
volume of such data, for example more than 10,000 tweets in the analysis of Twitter
data, it is no longer possible to skim or even read all the material. The initiating work
with texts and the exploration of the data must be limited to a specific selection. It is
quite natural that with these types of data there is a desire to integrate automated or
semi-automated steps into the analysis process.
In summary, QCA is suitable for all forms of analyses, for which systematic proce-
dures and methodological control are considered to be very important. As has been
made clear in the definitions above, categories play a central role in QCA. There are no
restrictions with regard to the data to be analysed, but the objective of systematics and
the comparison of data requires that the data must have a certain comparability.
Qualitative content analysis is therefore less suitable if only a small amount of data has
been collected scattered over a long period of time and is intended to record develop-
ments and changes. This is prototypically the case in ethnological field research, where
participant observation may take place again and again over a very long period of time
in order to find out about certain customs, rites, and rules. Here, however, QCA can be
used to evaluate certain sub-questions and sub-surveys, such as interviews conducted
specifically with a certain group in the context of field research.
1 The category-based approach and the key position of the categories for the
analysis
2 The systematic approach with clearly defined rules for each step of the analysis
3 The classification and categorization of the whole data set and not just a small
part of it.
However, a closer look reveals relevant differences even in these three points. For exam-
ple, with regard to the systematic approach, which in QCA includes several steps and
often also has a circular character, whereas it proceeds linearly in classical content analy-
sis. The decisive point, however, is that QCA is a type of analysis in which understanding
the meaning of a text and text interpretation play a much more important role than in
classical content analysis, which is focused on the manifest content. Authors of text-
books of classical content analysis such as Krippendorff (2018) and Früh (2017) argue
that the differences between classical content analysis and QCA are not so substantial
as to justify a dichotomy. In principle, we agree with this, but this does not include the
empirical examples presented by these two authors, such as the topic frequency analysis
presented in detail by Früh. The differences from this type of quantitative content anal-
ysis are significant.
The difference between qualitative and classical content analysis is particularly sig-
nificant for computerized automated content analysis, which aims exclusively at
statistical analysis, as it has developed since the mid-1960s, primarily in the USA. In this
type of analysis, words are coded automatically by means of a dictionary, whereby the
ambiguity of words and the question of meaning are largely ignored. In contrast, QCA
is an interpretive form of analysis, where coding is done on the basis of interpretation,
classification, and evaluation; the text evaluation and coding are thus linked to a
human performance of understanding and interpretation.
In conclusion, QCA differs in essential points from classical quantitative content
analysis:
• Fifth, the categories in QCA have above all a function of structuring and
systematizing and not of transforming data from the empirical into the world of
numbers.
• Sixth, QCA does not necessarily mean statistical data analysis. Statistical analyses
can take place within the framework of QCA, but unlike in classical content
analysis, statistics may play only a secondary role or QCA might do without
statistical analyses altogether.
approaching this world and its social problems. The term method comes from the Greek
word μέθοδος, meaning a pursuit of knowledge, an investigation, or a mode of prosecut-
ing such inquiry. A method is a tool for gaining insights, a planned procedure based on
rules, just as the principle of the diesel engine, for example, is a procedure for generating
energy through combustion so that a BMW can be powered. Methods can be applied in
very different contexts: internal combustion engines are also used in lawnmowers,
hedge trimmers, ships or aeroplanes. As a method, QCA can be used, for example, in the
context of grounded theory research or in the context of discourse-analytical research;
followers of the rational choice approach can also make use of it. Unlike variants of
discourse analysis, which have developed as methods out of Foucault’s discourse theory
(Keller, 2011), QCA has no background theory. This is by no means a shortcoming.
Quantitative methods such as variance analysis, multiple regression, and approaches of
causal modelling, such as the LISREL approach, do not have a background theory either,
but can be used in very different theoretical contexts. Sometimes there are close rela-
tionships between methods and theoretical approaches, such as between Bourdieu’s
theory of habitus and correspondence analysis (Blasius, 2001) but then the method of
correspondence analysis can also be used in the context of other theoretical approaches.
Another suggestion by Stamann et al. (2016, para. 11) is to make a conceptual separa-
tion of method and procedure. This does not seem reasonable to us. A quick look in any
encyclopaedia or dictionary shows that the two terms are rather to be understood as
synonyms. For example, Wikipedia states that ‘In recent centuries it [method] more
often means a prescribed process for completing a task’ (accessed 27 May 2022).
and effectively, but its results can also be integrated excellently into so-called ‘joint
displays’ (Guetterman, Creswell, & Kuckartz, 2015; Guetterman & Fetters, 2022). In
Kuckartz and Rädiker (2021), we describe numerous integration strategies for presenting
qualitative and quantitative data and/or results of a mixed methods study in such joint
displays. In practice, there are numerous joint displays in which typical results of
qualitative content analyses (identified themes, case-by-case thematic summaries, etc.)
are presented (Guetterman, Fetters, & Creswell, 2015). The frequent use of QCA in
mixed methods projects is also due to the fact that the results of qualitative analysis can
be transformed into quantitative data (‘quantitizing’) and analysed in combination
with the quantitative data.
for thinking. At the level of the researcher, the question of openness presents itself dif-
ferently, namely as openness in terms of communication during data collection and
(theoretical) prior knowledge. Does the postulate of openness here mean that the
researchers are not allowed to ask any specific questions? Does it mean, for example,
that in an open-ended interview, the researchers have to wait for narratives from the
research participants after an initiating stimulus and are not allowed to ask any further
questions? Or does the postulate of openness in relation to the researchers mean that
they should approach the research field and the research participants without prior
knowledge? As a consequence, this would mean that researchers should go into the field
as a tabula rasa, and not read any literature beforehand, but should engage in the
research situation without bias and not blinded by existing theories.
These questions can only be answered from the perspective of the chosen research
approach and the design of the research project. Qualitative content analysis, it should
be emphasized again, is not a research style, methodology, and certainly not a paradigm
according to which research is shaped and research designs are conceived. It is a method
of analysis and as such can be used in many research contexts and disciplines. It can be
used to analyse the material by means of categories; these can be defined in advance or
developed directly based on the material, whereby mixed forms are possible. In this
respect, the openness of QCA may vary, depending on the research approach chosen. A
characteristic of QCA is a high degree of flexibility; it can work with a category system
derived from theory as well as with categories developed entirely from the material.
Qualitative content analysis can be used to analyse very open interviews, such as narra-
tive interviews, as well as problem-centred interviews and strongly structured answers
to open questions in surveys.
citizens in Germany, underwear is usually changed daily, but pyjamas are only changed
every few days? Modern personal assistance software such as Apple’s Siri, Cortana
(Microsoft) or Alexa (Amazon) also formulate answers in natural language without them-
selves associating any subjective meaning with them, unless this software has already
taken on a life of its own behind the developers’ backs and is now producing meaning.
As you can see, we quickly get into a philosophical discussion that would transcend the
question of understanding meaning in the context of the method of QCA.
The term ‘sense’ has many meanings; for example, it can mean the senses of our
perception or the inner relationship of a person to something (‘I have no sense of clas-
sical music’). But secondly, according to Wikipedia, it can also mean the meanings and
ideas associated with a linguistic expression and thirdly, the state, the orientation of a
person’s thoughts. The above question about understanding meaning in the context of
QCA presumably aims to understand the meaning of a linguistic expression. But perhaps it
also means understanding the state and thoughts of a person. However, these are two fun-
damentally different orientations of understanding meaning. To put it clearly, QCA is
not really well suited to the latter, it does not want to and cannot understand the inner
states of a person; this is the professional field of psychology, and understanding mean-
ing in this way is probably not easily practised by people without psychological training.
Moreover, everyday life shows that it is very difficult to understand other people’s state
of mind in a valid or real way. Qualitative content analysis is not a method of individual
psychological understanding of meaning, in terms of empathizing with inner worlds.
Projects that work with QCA are mostly concerned with understanding the meaning of
social action and understanding the meaning of statements and arguments. How the
interpretive understanding and explanation of social action is to take place needs to be
decided before QCA starts. Researchers can, for example, use Max Weber’s widely known
types of action (Prosch & Abraham, 2006) or Esser’s (2001) framing model as a guide. As
we have described in Section 1.4, we believe that hermeneutic techniques can provide
important guidance. With the method of QCA, no model of understanding is mandated
here, but it is required that the researchers develop a category system with consistency
of meaning on the application of the categories when coding the material.
We think it is important for researchers to know that QCA has long history as a
method of social research. As a classic, predominantly quantitative content analysis,
QCA emerged as early as the first half of the twentieth century. For more than 60 years,
researchers have now used the term QCA to describe a decisively qualitative method.
We see an important connection between QCA and principles of text understanding,
such as those developed in hermeneutics: to understand the meaning of a text as a
whole, one must understand the meaning of its parts – and vice versa. We conceptualize
QCA as a qualitative method with a systematic approach and a claim to produce inter-
subjectively valid results. In this regard, it differs from strongly constructivist
approaches.
As a method of analysis, QCA is not in competition with methodologies and research
styles such as grounded theory methodology or critical-rational approaches. Therefore,
QCA can also be used within the framework of various overarching approaches such as
the documentary method, discourse analysis, or even grounded theory.
Chapter objectives
In this chapter you will learn about:
Content analysis stands or falls by its categories … since the categories contain
the substance of the investigation, a content analysis can be no better than its
system of categories. (Berelson, 1952, p. 147)
In principle, this statement also applies to QCA. Here, too, (almost) everything revolves
around categories and so it is obvious to pay special attention to the term ‘category’.
1
https://en.wikipedia.org/w/index.php?title=Categorization&oldid=1023744955
(accessed 2 August 2022).
for everyday life and decisions as well as for practising science because objects in the
world around us do not themselves dictate to which category or class they should be
assigned. Thus, we have to assign objects and ideas to categories, and our perceptions
and thought processes influence every categorization we make.
The question of what exactly a category represents in empirical research is hardly
addressed in the literature on research methods, even in textbooks that focus on methods
of qualitative data analysis. It is apparently assumed that people already know what a
category is based on common sense. Instead of a definition, you often find a collection of
category attributes, particularly in the context of the method of classical, quantitatively
oriented content analysis. For example, Pürer (2003, p. 551) states that text features
should be recorded ‘with the help of a systematically developed category system with
clearly defined categories. The categories must be clearly distinguishable from each other’.
Category Source
42 Energy consumption/energy Category with number 42 from a quantitative content
demand analysis (Früh, 2017, p. 169)
13100 Afghanistan conflict Category with number 13100 from a quantitative
content analysis (Rössler, 2017, p. 131)
Societal impact Category from a qualitative content analysis on climate
awareness
Baker Category for occupational classification
Protective governing Category developed in a grounded theory study
(Strauss & Corbin, 1990, p. 121)
Personal affectedness by climate Evaluative category developed in a study of social
change with the following values: science environmental research
– high level of affectedness
– low to medium level of
affectedness
– not affected
– not ascertainable
Scope (length/duration) Category in a media analysis (Rössler, 2017, p. 116)
Critique of multiform racism with Category developed in a grounded theory study in the
marking of resistance options research field of digital youth cultures (Dietrich & Mey,
between discourse orientation 2019)
and use of violence
Factual categories. These are categories that refer to certain objective or seemingly
objective occurrences, such as to classify different occupations (someone is a
‘politician’; someone says, ‘I am a baker’), a place (‘I live in Surrey’, ‘I live in a
redevelopment area’) or an event (‘train accident at Rome Central Station’).
Evaluative, scaling categories. With the help of these categories, assessments of specific
data are made on the basis of the researchers’ evaluation standards. Evaluative
categories have a defined number of values that are used to assess the data. Evaluative
categories often have an ordinal scale level; for example, the category ‘helper
syndrome’ may have three values, namely ‘strong’, ‘little’, ‘none’. Dichotomous
rating scales (a characteristic is ‘present’ or ‘not present’) can also be used. Coders
process relevant passages of the material and make a rule-based classification.
Analytical categories. This type of category is the result of the researchers’ intensive
examination of the data, that is, the categories move away from the description
that is provided by thematic categories. For example, an analysis of the thematic
category ‘environmental behaviour’ and its dimensions ‘mobility behaviour’,
‘energy behaviour’, etc. leads the researchers to the realization that the research
participants often talk about the financial costs and benefits of certain behaviours;
they then define the analytical category ‘cost–benefit calculation’.
Natural categories. These are the terminology and concepts used by the actors in the
field themselves. In the English literature on research methods, especially in
grounded theory, the term ‘in vivo code’ is used for this purpose (Strauss & Corbin,
1990, p. 69). An example of this is the term ‘tradition bearer of the unit’ mentioned
by Strauss, by which a head nurse refers to another ward nurse (Strauss & Corbin,
1990, p. 69). The transition to analytical categories is fluid because actors use these
terms to explain to themselves and others the phenomena of their everyday world.
Often natural categories are very vivid and pictorial, for example, when young
people, during interviews, refer to a particular teacher as ‘the eco-woman’.
Formal categories. This category type refers to dates and information about the unit
being analysed. For an open interview, for example, this could be the length of the
interview in minutes, the date of the interview, and the number of words.
In qualitative content analysis, another form of categories often plays a role, the so-called
ordering categories. These are not used to code the material, but to structure the categories
themselves. In this way, the ordering categories have a similar function to the headings
in a book or in an article. This considerably improves the clarity.
We see the importance of such a distinction of category types primarily in the fact
that it promotes analytical sensitivity in a qualitative content analysis. Furthermore, it
initiates a probing and comparative reflection on the appropriateness of category types
for the respective research questions. Although in principle it is always a matter of
assigning a category to a certain segment of the material, for example a certain passage
of text, the processes of coding differ depending on the type of category. This soon
becomes clear when you compare the use of natural categories and factual categories:
in the one case, one searches for particularly apt ways of speaking by research partici-
pants in an exploratory and interpretive way; in the other case, it is about identifying
facts, features, and attributes that are directly related to the research questions.
Category types and the forms of coding are thus intimately connected. While we
approach the relationship of categories and coding from the side of the categories and
distinguish category types, Saldaña (2013) takes the perspective from the side of the
coder. He distinguishes numerous types of coding, including ‘attribute coding’ and ‘in
vivo coding’ corresponding to the ‘factual categories’ and ‘natural categories’ examples
above. Saldaña’s almost encyclopaedic presentation of coding forms in research prac-
tice stimulates the analytical imagination and is especially helpful if one wants to
know in which studies certain types of coding, for example, ‘narrative coding’, ‘emo-
tion coding’, or ‘dramaturgical coding’, were practised. Saldaña distinguishes 32 types
of coding and invites readers to fill in other coding forms they encounter in the litera-
ture or discover for themselves in the blank forms included in his book (Saldaña, 2013,
p. 183). Whether it is really useful to distinguish between so many different types of
coding for QCA work seems questionable, especially because, on closer inspection,
many types of coding overlap and because they are located at completely different
levels of abstraction.
The confusion increases even more because in qualitative social research the term
‘code’ and corresponding specifications (sub-code, higher-level code, parent code, top-
level code, etc.) are also frequently used (e.g., in Bazeley, 2021, pp. 155–260; Miles et al.,
2020, pp. 61–102). This applies first and foremost to grounded theory, where the term
‘code’ appears in several forms, namely as ‘open’ and ‘axial code’ in the early stages of
analysis and as ‘substantive code’, ‘key code’, ‘selective code’, or ‘theoretical code’ in the
later phases of analysis (Strauss & Corbin, 1990, pp. 57–142). Corresponding to the dif-
ferent analysis activities in the different phases of the analysis process, the term ‘code’
sometimes denotes an abstract concept encompassing many dimensions, but some-
times only a first ad-hoc developed concept, which possibly develops into a category
within the further analysis. This now introduces another term, namely ‘concept’, whose
relation to the terms ‘code’ and ‘category’ is also not quite clear. Although Strauss and
Corbin (1990) have endeavoured to provide some clarity regarding the different uses of
the terms, it is difficult to deny that although some grounded theory writings differen-
tiate between the terms ‘code’ and ‘category’, often their use is not uniform and
consistent. This applies, for example, to the various editions of the textbook by Strauss
and Corbin. In addition, scholars advocating grounded theory usually work with an
expanded concept of coding, which includes analysing, naming, and categorizing as
well as the theoretical classification of the data as a whole.
Obviously, there are different traditions and conceptual cultures in qualitative
research, and also within grounded theory itself. For example, the constructivist variant,
as represented by Charmaz (2014), uses some of the same terms as Glaser (2005) and
Corbin and Strauss (2015) but also others, such as ‘focused code’. Overall, in the field of
qualitative data analysis, one encounters numerous inconsistencies and confusing phe-
nomena in relation to the use of the terms ‘code’ and ‘category’. Thus, with regard to
the activity, one speaks almost always of coding, but far less often of categorizing.
Moreover, in translations from English into other languages, ‘code’ is often translated
in the same way as ‘category’.
A look beyond qualitative research shows that in communication studies and the
prevailing form of content analysis there, people usually talk about categories and not
codes. However, the application of categories to research data, such as media products, is
referred to as coding and not categorizing, and the people who do this are called coders
(Früh, 2017; Krippendorff, 2018). Those who are at home in quantitatively oriented social
science research will be particularly surprised, because code, category, and coding mean
something quite different there than they do in the context of qualitative data analysis:
there, ‘code’ refers to the assignment of a number, or to put it more generally, a sign, to
a certain attribute (also called a ‘category’). For example, in a data matrix, respondents
from rural areas can be coded ‘1’ and respondents from urban areas ‘2’. For them, coding
is therefore a transformation process from the empirical world to the numerical world.
Of particular importance is that in qualitative data analysis software (see Chapter 8),
the term ‘code’ is used almost exclusively, whether it is a code, a category, a main or key
category, a concept, or a theme.
All this makes efforts to delineate the terms appear very difficult and unpromising.
Even if it were possible in this book, readers would be confused as soon as they picked
up other relevant books. The question may therefore be asked whether all the terms
mentioned do not ultimately mean the same thing and only a blurred use of terms
needs to be stated. This can certainly be said in the affirmative for some methodological
texts and research approaches. As described, particularly in the literature on grounded
theory, there is often no systematic, continuous distinction between concept, code, and
category. There is therefore a case for using the terms ‘code’ and ‘category’ synony-
mously, including the fact that some composites are more common than others in
English (e.g., ‘coding frame’ instead of ‘category frame’ or ‘categorizing frame’, and
‘code memo’ rather than ‘category memo’), while in other languages, such as German,
the opposite might apply. One argument against such synonymous use is that there are
plausible efforts to ascribe differentiated meanings to the terms and to establish a
chronological order oriented to the research process. For example, Creswell and Creswell
Báez (2021, pp. 158–171) describe the path from the raw data (texts) to the themes in
the following sequence: first, read texts and assign codes to text segments; second, order
codes and summarize codes into themes; third, create a ‘map of the themes’ and
describe their interrelations in the ‘qualitative report’. In this inductive approach, the
initial, still poorly developed, codes become themes in the course of the analysis, or
categories in the language of qualitative content analysis. Such a development from
simple to more abstract and complex analytical tools is typical of many analytical
approaches that work with codes or categories, and that is why we also want to recom-
mend it for qualitative content analysis. Thus, in inductive approaches (what exactly
these are is described in Chapter 3), as far as the beginnings of the analysis are con-
cerned, we speak of codes. But when it comes to the main phases of qualitative content
analysis, we should always speak of categories and not of codes. The title of Chapter 3
of this book is therefore ‘Developing Categories’ and not ‘Developing Codes and
Concepts’. We will try to maintain a consistent use of terms in relation to QCA.
However, in so far as we refer to grounded theory or positions close to it in the rest of
the text (such as the technique of open coding), we will not attempt to impose a consist-
ent use of terms on grounded theory ex post and will not attempt to introduce a kind
of standard language. The same applies to Chapter 8, in which we describe the imple-
mentation of qualitative content analysis with QDA software. In the current QDA
programs, only the term ‘code’ and its composites (e.g., ‘sub-code’, ‘code name’, etc.)
exist – we cannot change this and so we naturally also use these terms, which some-
times seems a little strange, for example, when we describe how new categories are
inserted with the option ‘new code’.
As far as the relationship of the terms ‘theme’ and ‘category’ is concerned, we under-
stand these terms to be synonyms if they are thematic categories. For example, when
summarizing the coded text segments of the category ‘personal environmental behav-
iour’, we would speak of a thematic analysis rather than a categorical analysis.
We understand the term ‘concept’ as transversal to the terms code, category, and
theme. A concept can be both a designation for individual events, phenomena, and
occurrences in the first phase of the analysis (Strauss & Corbin, 1990, pp. 61–69) as well
as an elaborated idea, a complex plan, for example, the ‘sustainability concept of the
Green Party’ (resulting from the analysis of documents) or ‘Heidi Klum’s concept of
beauty’ (resulting from the analysis of various episodes of Germany’s Next Topmodel). In
this sense, a concept can be a code, a theme, or a key category that emerged through
analysis only at a late stage of the research.
The third type of category system organization, the network structure, is characterized
by the fact that the elements (nodes) of the network can be connected to each other in
many ways (and not only hierarchically). Networks are usually represented as graphs
with nodes and edges. Networks allow different connection paths, a fact that is used, for
example, in the organization of websites.
The category systems of qualitative content analysis are almost always structured
hierarchically. This also applies to the examples presented in this book. Compared to
linear lists, hierarchical category systems have the advantage that they can be struc-
tured, and the groupings can convey meaning. This suits a multi-level approach that
wants to go into detail and discover dimensions and specifics through the analysis.
Compared to network-like category systems, hierarchical category systems are clearer,
and they allow one to search for connections at different levels. The advantage of hier-
archical structuring may be illustrated by the following example of geographical
structuring. With appropriate structuring, we can compare countries with each other
(e.g., Germany and Canada), regions of these countries (Bavaria and Quebec), cities of
these regions (Munich and Montreal), certain neighbourhoods in these cities, and so on.
It is easy to aggregate and disaggregate, an advantage that is particularly useful when
using QDA software.
Are there certain demands to be made on category systems? The question can be
answered with a clear ‘yes’. Category systems should have an inner coherence and their
structure should be plausible and comprehensible. The main categories should share a
similar degree of abstraction and generality.
Furthermore, how can you recognize a good and useful category system and how
does it differ from a less good one? The following standards should be met (according
to Rädiker & Kuckartz, 2020, pp. 46–49):
The categories are closely related to the research questions. This is the most important
condition; after all, the categories should contribute to answering your research
questions. It is therefore important to ask of each category to what extent it helps
to achieve the goal of a study. The usefulness of a category can, of course, also
consist in capturing important contextual knowledge, that is to say, a category does
not have to be removed just because the connection to the research question is not
immediately obvious. The examining gaze should also grasp the overall construct:
Does the category system offer sufficient analytical depth? It can be helpful to
anticipate the finished results report and its structure: does the category system
cover all important aspects to be investigated or are categories relevant to the
research questions possibly missing?
The categories are exhaustive. This means that for every aspect in the data that is
important for answering the research questions, and hence should be covered,
there is a category. Whether this criterion is met can only be said with certainty
when the categories are applied to the empirical data, (i.e., it is primarily an
empirical question). Often, at the beginning of the analysis, it is a good idea to
The categories are selective (distinct) but, depending on the application, they can also be
mutually exclusive. ‘Selective’ means that it must always be clear which category is
assigned to a particular section of text – and which is not. If you constantly
fluctuate between two similar categories when assigning categories, this is an
indication of low selectivity. Absolute unambiguity cannot always be achieved;
nevertheless, special attention should be paid to precise category definitions.
Precise definition of categories, however, does not mean that only one category
may always be assigned to a text passage. Very often in the application of QCA,
several main categories or sub-categories are assigned to the same text section,
sometimes even overlapping, because several aspects are addressed. In some cases,
however, the multiple assignment of categories is not desired, and it is necessary
that categories are mutually exclusive. In these cases, categories are not allowed to
overlap. Typical examples are evaluative sub-categories, such as ‘low’, ‘medium’,
and ‘high’ pro-environmental attitudes.
The categories are well formulated. It is important to pay attention to how category
names are formulated. Are equivalent sub-categories also formulated linguistically
in an equivalent way? It also makes a difference whether you choose ‘climate’ or
‘climate crisis’, ‘motivation’ or ‘attitude’ as a category. It is helpful to look at a
dictionary or thesaurus to include distinctions from other terms in the category
definition and to select terms that are as appropriate as possible.
Taken together, the categories form a ‘gestalt’. We have already addressed this aspect in
the introductory sentences above. The point is that a category system has an inner
coherence and that the categories do not merely stand next to each other loosely
and without connection. To check this criterion, it is helpful to pay attention to
the level of abstraction of the categories; in particular, equivalent sub-categories
should have a comparable degree of generality. The main categories should also
generally have a similar level of abstraction and generality.
The sub-categories are dimensions, values, or sub-aspects of their parent category. The
degree of abstraction sensibly decreases with each lower hierarchy level, since the
sub-categories should always be aspects, values, or dimensions of the parent
categories. While this seems logical, it can entail a longer development process in
category development on the material. Is, for example, ‘nature’ mentioned in the
interview as a sub-aspect of ‘environment’ or is it on the same level so that it
would be better called ‘nature and environment’? Should ‘climate change’ be a
separate main category or a sub-category of ‘environmental issues’?
The categories are understandable. The category names should not be too
complicated. Names should be used that are easy to understand for the coders as
well as for the later recipients of the study. This does not mean that no technical
terms should be used and category names should be formulated in simple
language. However, when formulating and constructing the category system, it is
useful to keep in mind who will be working with the categories and to whom they
will later be presented in publications, lectures, and posters.
A category definition should have the structure shown in Figure 2.1. Quotations
from the texts should help to illustrate the concrete application of a category, that is,
examples must be selected that are as typical as possible for a category. It is not always
possible to find a truly typical example that covers all use cases of a category. In this
case, several quotations can be integrated into the category definition, or a typical
example can be constructed as an alternative or supplementary example, which must of
course be marked as such.
Category definitions have a twofold function. First, they document basic elements of
QCA for the recipients of the study and for the scientific community as a whole.
Without knowledge of these basic elements, the results of the analysis may be difficult
for outsiders to interpret. Second, the category definitions – supplemented by concrete
instructions for action – represent the coding guidelines for the coders. This means that
the more accurate the definitions are, and the more illustrative the examples of the
applications of each category are, the easier it is to code and the more likely it is for a
high level of agreement among coders to be achieved.
The category manual is to be distinguished from the coding guide; alternatively, the term
code book is also used. The term ‘code book’ has nothing to do with secret services or with
cryptology but refers to a document that contains all categories and their definitions.
The difference between the coding guide and the category manual can be expressed
most simply by an equation: category manual + instructions for coders = coding guide.
The coding guide is a document that is primarily intended for internal use and gives
coders specific instructions for their work. The code book, on the other hand, is
designed for the world outside of the research project. Categories are central to QCA,
and in this respect the code book has a very important function, as it documents the
accuracy and precision with which the work was done.
The result of both views is ultimately the same, namely a connection between the
text passage and the category. Figure 2.2 illustrates this basic principle: on the left is the
original text, on the right the coded segment. The section with a grey background was
coded with the category ‘learning via values education’.
Figure 2.2 The coding process: original text, category, and coded segment
Before you can get started on trying out your coding frame, you first have to
divide your material up into smaller units, which you will then code using
your coding frame. This is called segmentation. (Schreier, 2012, p. 126)
In QCA, we believe that coding has a broader meaning than mere indexing or ‘tagging’
and that a predetermination of coding units is not necessary but rather alien to the
process of the analysis. In the context of QCA, researchers usually code units of mean-
ing instead of determining coding units in advance, which means that coded segments
may well overlap or be nested within each other. The criterion for determining the seg-
ment boundaries is comprehensibility ‘out of context’, which means that the segments
should be comprehensible in themselves when seen outside their context. Of course, it
does not necessarily have to be units of meaning that are chosen as coded segments. In
the case of factual categories or natural categories (in vivo codes), the situation is differ-
ent; also, the occurrence of certain persons or places, certain metaphors or idioms can
be coded, that is, in this case the coded segments are very short and do not represent
units of meaning.
Often, especially when the amount of text to be processed is very large, assistants
beyond the scientific research team are called in specifically for the coding of the mate-
rial. For QCA, a certain degree of interpretive competence is necessary: the coders must
be well informed about the research questions, the theoretical constructs and the mean-
ing of the categories. Typically, team coding sessions and coding workshops are
conducted to achieve as much coding agreement as possible. While in quantitatively
oriented content analysis appropriate coefficients of agreement, such as Krippendorff’s
alpha, Cohen’s kappa, or Scott’s pi (Krippendorff, 2004b) are calculated to determine the
so-called intercoder reliability or interrater reliability, in QCA there is a tendency to
adopt a procedural approach that seeks to minimize non-matches through discussion
and consensual decision-making within the research team. Chapter 9 describes this
approach and the options for determining intercoder agreement in more detail.
Chapter objectives
In this chapter you will learn about:
If you choose to conduct a qualitative data analysis, you will likely ask yourself, ‘How do
I determine which categories to use?’, ‘How many categories are necessary?’, or ‘What
rules do I have to follow in constructing the categories?’
In the literature on research methods, little information is given about how exactly
categories are constructed because it is assumed to follow common sense. You may
come across rather unhelpful statements, such as that there are no patent remedies for
developing categories (Kriz & Lisch, 1988, p. 134). However, the same textbooks point
out that category construction is very relevant for the analysis and protagonists argue
that content analysis stands or falls with its categories. So, you may be asking yourself,
how in the world can you construct something as important as a category?
The most suitable way to develop the categories depends largely on the research
questions at hand and any previous knowledge that researchers have about the given
research subject or field. The more theory-oriented the project, the more extensive the
previous knowledge, the more focused the research question, and the more specific the
existing hypotheses, the easier it is to develop categories while reading through the col-
lected data.
Looking at empirical studies, a spectrum of theoretical and empirical category
development can be observed:
Theory. If a theory guides a study, the constructs and concepts contained in the
theory can be translated into categories. For example, if an interview study examines
media literacy in assessing (fake) news on climate change, the interview will certainly
not directly ask about this media literacy. Nevertheless, the theoretical elaborations
on media literacy should be consulted in the category development to ensure that
important theory-based constructs are accessible in the analysis. The important
constructs of the guiding theory are usually also found in the research questions, that
is, there is a triad with the three interrelated elements of theory, research questions,
and categories that need to be reconciled in the development of deductive categories.
State of research. It also makes sense to consider the current state of research during
the development of categories. If connections between media literacy and the
assessment of (fake) news have already been empirically verified, these can be
incorporated into the category development. There may even be existing category
systems of thematically similar studies that can at least be adopted as a starting point.
Everyday knowledge. There is nothing wrong with using everyday knowledge to develop
deductive categories as long as the result contributes to answering the research
questions in a meaningful way. If the aim is to investigate why young adults drop out
of university, or what motivates students to spend a semester abroad, categories can be
developed a priori on the basis of everyday knowledge without reference to theory or
collected data. The categories developed in this way can supplement theory-based
categories or be compared with the result of a later inductive category development.
These sources of deductive categories are not mutually exclusive. On the contrary, differ-
ent sources of deductive categories should always be considered in a research project to
increase the chance of including valuable and innovative categories.
1 Politics
2 Business
3 Finance
4 Sport
5 Local
6 Culture
7 Miscellaneous
These seven categories were developed by the newspaper editors independently of the
data (i.e., the news items). The different categories seem plausible because they are based
on everyday knowledge and the perceived social realities in our culture, in which these
categorical differentiations can be found in scientific disciplines and governmental
agencies. Another newspaper can of course work with completely different categories
such as ‘Technology & Motor’ and ‘Life’. Now, any new story that appears on the news
ticker can be ‘coded’ accordingly and forwarded to the appropriate editor or division, as
can be seen in Figure 3.1.
Assigned
department/
Time Headline category
18:48 Buried victims freed by companions Miscellaneous
18:24 EU debt limits remain suspended in 2022 Finance
17:58 Spectacular prison break in Athens Miscellaneous
17:22 Will Israel’s coalition be sworn in as early as Wednesday? Politics
16:17 Another avalanche in the Alps Miscellaneous
16:05 World’s largest meat company ramps up production again Business
15:52 The giant batteries from Wittenberg Business
15:16 NASDAQ at lowest point since autumn 2004 Finance
15:10 David Diop wins British International Booker Prize for Literature Culture
15:08 Why laziness is important for democracy Miscellaneous
Clearly, some stories or topics are difficult to assign. For example, it is not clear if the
story with the headline ‘NASDAQ at lowest point since autumn 2004’ should be
assigned to the ‘business’ or ‘finance’ category. Thus, it is imperative to formulate crite-
ria to be able to differentiate between the categories and encompass the intentions
behind the classifications so that coders can reliably assign the headlines and code the
data. In content analysis, such a written assignment rule is called a category definition. In
our example, such a definition could be as shown in Figure 3.2.
A category definition must at least contain the name of the category and a descrip-
tion of its content. Furthermore, it is very useful if concrete examples and delimitations
to neighbouring categories are included. When applying the categories, concrete exam-
ples from the data should be added.
The biggest difficulty in constructing categories deductively lies in formulating pre-
cise category definitions so that the categories do not overlap. Furthermore, there is the
need for completeness of the categories. For example, the category scheme with seven
categories presented above could hardly be used if we forgot to include a category like
‘business’. Also, it is always important to include a category for topics that may not fit
any of the other categories (‘miscellaneous’ here); this will allow you to assign all of the
data. In the methodological literature, the requirement for categories in deductive cat-
egory development is therefore often that categories should be mutually exclusive and
exhaustive (e.g., Diekmann, 2007, p. 589; Krippendorff, 2018, pp. 138–139). In the
above example this means that a news item from the news ticker should not be for-
warded to two departments at the same time, otherwise one would risk the same story
appearing in several places in the newspaper. In a newspaper, it does not make sense to
run a news item twice. In QCA, on the other hand, a statement can also contain several
aspects and therefore be assigned to several categories if necessary. Only by defining the
categories as precisely as possible can sufficient quality be achieved in the application
of the categories.
Table 3.1 shows how categories can be generated from an interview guide on climate
awareness. The category names refer to the main focus of the interview questions and
summarize it in one or more words, taking into account the research questions. The names
can be very short and concise, as in the example, which is usually very practical for further
presentation and use in QDA software. However, they can also be formulated in more
detail, as we describe in Section 5.4. For example, instead of ‘Learnability’, ‘Learnability of
dealing with global problems’ could also be used as a category name. In any case, it is neces-
sary to record the meaning of the categories in category definitions as far as possible a priori.
The task for the small groups was to develop a category system of thematic categories
for a QCA in advance (i.e., without looking at any of the answers). The categories were
to build on the state of research on ‘quality of life’. For this purpose, the workshop par-
ticipants were asked to explore the state of research on the internet – for example, in
the online encyclopaedia Wikipedia.
Figure 3.3 shows the category systems of seven different working groups (with three
or four members each) that emerged in the course of these exercises. Comparing the
category systems developed by the seven working groups, the following observations
can be made:
1 The number of categories proposed by the groups varies; it ranges between five
and eight categories (for pragmatic reasons a limit of 10 had been set).
2 It is noticeable that no two proposals are identical; however, there is sometimes a
great similarity between different proposals – for example, between the category
proposals of groups 1 and 3.
3 Not all category systems are exhaustive; for example, proposal number 7
obviously lacks important aspects of quality of life.
4 Only one thematic category, ‘Health’, is proposed by all groups.
5 A number of categories appear in several proposed category systems (e.g., ‘education’,
‘social relations’, ‘work and occupation’) but with quite different accentuations (e.g.,
‘education’, ‘educational offerings’, ‘opportunities for education’, etc.).
6 There are proposals where the categories have obvious overlaps and therefore
raise the question of principle delimitations; for example, for group 6, it may be
unclear how the statement ‘material security’ would be coded if there are the
categories ‘standard of living’ and ‘security’.
7 The proposed categories have different degrees of abstraction or generality; for
example, ‘Social integration’ is much more general than ‘Family & friends’.
In another workshop we combined the seven proposals into a single category system for a
QCA. How did we proceed? Firstly, we followed the principle that frequently occurring cat-
egories are taken over into the common category system. Then we discussed in detail the
delimitations between the categories and added definitions to the individual categories to
ensure selectivity. The result was the following category system consisting of 11 categories:
In the process of creating the category system, decisions often have to be made that have
a far-reaching impact on the subsequent analysis and the results of the research proj-
ect. In case of doubt, the research questions should always be consulted as a decision-
making aid and should be the deciding factor. At this stage very specific categories, such
as ‘Leisure activities’ or ‘Theatre’ should be avoided, and instead more general categories,
such as ‘Culture’ should be preferred, which include the more concrete aspects. In order
to cover everything that could be mentioned about quality of life, a residual category
‘other’ should normally be included. If it turns out later that a large number of state-
ments are coded with this residual category, it will be possible to add another category if
there is an accumulation of certain aspects.
It is important that the categories are exhaustive and do not overlook a theme that
is important from the perspective of the research questions. The subsumption under
more abstract and general categories can have implications that can trigger criticism and
contradiction in the later reception of the study in the scientific community. For exam-
ple, the grouping of ‘Nature’, ‘Environment’, and ‘Sustainability’ under one umbrella
category is not unproblematic; protagonists of the sustainability model emphasize pre-
cisely that this is not just an environmental concept, but that the model also includes
social and economic dimensions. Furthermore, one should think ahead to the later work
of coding the data and the desired agreement of the coders; a broad interpretation of
sustainability could lead to a very, very large number of statements being identified as
belonging to this category and the categories losing the required selectivity.
With the later communication of the results of the QCA in mind, it can also be very
important to integrate new thematic aspects and to develop inventive categories that
may, for example, not be as obvious as ‘Health’. In the category system above, the
category ‘work–life balance, time prosperity’ is an example of such an inventive cate-
gory. These types of categories signal a new approach to the topic, they are innovative,
and can advance research. In addition, one should also keep an eye on the communi-
cation and presentation of the study in the scientific community: in order to attract
adequate attention in today’s polyphonic scientific communication, a certain degree
of originality is absolutely necessary.
Can a judgement be made as to which of the seven proposed category systems is ‘bet-
ter’ and which is ‘worse’? Yes – using the criteria presented in Section 2.4. It is relatively
easy to determine whether a category system is exhaustive or not. According to this
criterion, the proposal of group 7 is certainly worse than the proposal of group 1. A
second important criterion for assessing a category system is the quality of the category
definitions (cf. Section 2.5). In this context, the quality criterion of coder agreement –
both intracoder agreement and intercoder agreement – comes into play as a third
criterion. This is a highly important criterion for assessing the quality of any form of
content analysis. If the categories formed are not selective, only a low level of agreement
can be achieved. Fourth, a criterion for a good category system that should not be under-
estimated but is difficult to operationalize is the coherence and plausibility of the
overall structure of the category system. There is a need to develop a plausible whole and
not just individual (potentially even selective) categories that stand next to each other
rather unrelatedly.
fulfils all the criteria of qualitative research such as openness, communicativity, etc.
(Hopf & Schmidt, 1993; Hopf et al., 1995; Hopf, 2016). This means that we are dealing
with a combination of deductive category development and qualitative research.
When using deductively developed categories, it may turn out that categories are not
selective or that very many passages are coded into the residual category ‘other’. This
leads to categories being modified or even to new categories being defined. Deductive
category development therefore in no way excludes the possibility of changes being
made to the category system and the category definitions during the analysis and thus
deviating from strict adherence to the a priori definitions.
For this reason, it does not make sense to calculate coefficients of agreement or intercoder
reliability in general for the development of categories. In many workshops we have prac-
tised getting the participants to form categories from the material individually and/or in
groups. There are always areas where the proposed category systems overlap (i.e., the same
or very similar categories are proposed). But just as regularly there are non-overlapping
areas and it is not uncommon that category systems are formed that are structured com-
pletely differently, are partly also very original and definitely represent a reasonable
implementation of the research question.
Where can we find good examples of how categories have been generated directly
from empirical data? Where have scholars given more detailed methodological reflec-
tions on the inductive development of categories?
This section first presents two approaches that have dealt with the inductive develop-
ment of categories: Mayring’s (2014, 2021) approach, which is based on paraphrasing
and summarizing, and the multi-stage procedure of grounded theory, which begins with
open coding and organizes the open codes into categories (Strauss & Corbin, 1990;
Charmaz, 2014). We then present guidelines for inductive category development, which
can be used as a guide for your own projects.
This example not only illustrates the general procedure, but also shows problems of
paraphrasing. For example, whether the generalization ‘No practice shock experienced …’
really reflects the original statement of the respondent seems rather questionable.
Anyone who reads the relevant pages in Mayring’s text (2014, pp. 70–76) will come across
many paraphrases and generalizations that seem rather problematic. At this point, how-
ever, problems of paraphrasing will not be discussed further and we will only focus on
the subsequent procedure up to the development of the categories. After paraphrasing the
text, a step described by Mayring as ‘reduction’ is carried out, with the help of deletions
and bundling of paraphrases, to create a system of statements that summarizes each indi-
vidual interview. The next steps of category development are then cross-case: the
categories of all interviews previously paraphrased are again bundled and summarized in
a second round of generalization and reduction, whereby Mayring forms the following
four categories in the example presented:
K′4 Wanting to try out pedagogical behavior strategies and still remaining consistent
in one’s treatment of the class presents a dilemma (Mayring, 2014, pp. 76–77)
The categories developed in this way are declarative sentences that are supposed to con-
tain the core content of the material in a condensed form. It remains relatively unclear
how contradictory statements of respondents can be reduced and summarized. The cat-
egories developed by continuous generalization apparently claim to be free of contra-
dictions. Here is an example. Two of the four trainee teachers interviewed in the sample
study reported great difficulties in dealing with students. However, in the category ‘K′3 A
good relationship with students can always be attained’, which was developed through
the summary procedure, nothing of this can be seen any more.
Another problem concerns the relationships of the sub-statements (sub-categories) to
the main statement (main category). In the case of category K′1, five factors are men-
tioned as conditions for ensuring that trainee teachers do not experience practice shock.
In the simplest case, assuming dichotomies (e.g., teaching experience present versus
teaching experience not present), results in 25 = 32 possible combinations of these five
factors. In which of these constellations does practice shock not occur? Is it only in the
one case when all five positive factors are present?
Mayring has formulated more than 20 specific rules for these eight steps. An example
of this procedure can be found in Mayring (2014, pp. 83–87). This example illustrates
not so much the process of category formation itself, more the result and the form of
the categories developed. In this example, based on the research question and guided
by Lazarus’s theory of stress, firstly a category definition is formulated: ‘Stressful
1
Barney Glaser and Anselm Strauss founded the tradition of grounded theory in 1967
(Glaser & Strauss, 1967), which is not simply a method or an evaluation technique.
Strauss later said that grounded theory is a research style or a methodology and that at
the time, he and his co-author wanted to design an epistemological and political science
approach that deliberately and provocatively went against the leading behaviourist
research paradigm (Strauss interviewed by Legewie & Schervier-Legewie, 2004). Kelle sees
an ‘inductive misunderstanding of one’s self’ in the beginnings of the history of grounded
theory (Kelle, 2007, p. 32). Grounded theory developed in three main directions: Strauss
and Corbin; Glaser; and Charmaz, who gave grounded theory a constructivist flair.
directly from the data through a multifaceted and circular process. It should be noted that
the original textbooks of grounded theory speak almost exclusively of code and not of cat-
egory, but in more recent textbooks, such as Charmaz’s (2014), both terms are used.
Open coding represents the first step in working through the data. It is centred on
identifying and/or naming concepts. In grounded theory, concepts are labels or tags for
phenomena, and they serve as the foundation for the theory to be generated. Strauss
and Corbin (1996, pp. 43–46) name the following examples of concepts: ‘attention’,
‘transfer information’, ‘offer support’, ‘monitor’, ‘satisfaction of guests’, and ‘experi-
ence’. In grounded theory, coding refers to the intellectual processing of empirical data
and the development and assignment of codes, where a code is a label that, to a certain
extent, theoretically sums up a segment of the data:
Furthermore, in the grounded theory approach – in contrast to QCA – the entire process
of data analysis is conceived as coding. Concepts have a similar role to that in standard-
ized quantitative research. Specifying concepts requires you to step away from the data
and work towards developing theories. Strauss and Corbin also refer to coding as diving
into the data. You can proceed line by line, or consider sentences, paragraphs, or even
complete texts, asking yourself what is the main idea of this sentence, paragraph, or text.
Researchers then name phenomena, pose questions, and make comparisons regarding
the similarities and differences that can be found within the data. How open coding
works in the style of grounded theory can be well illustrated by an example from Char-
maz. It is an interview excerpt with a very ill patient Bonnie (48 years, living alone),
who talks about how her daughter Amy only found out about her current state of illness
through her neighbour Linda:
She found out from Linda that I was, had been in bed for days and she called
me up, ‘You never tell me, and I have to find out from Linda’, and ‘Why don’t
you tell me who you are and what’s going on and …’ Well, I don’t know how
long after that, but that Saturday the pain started right here and it, throughout
the day it got worse and worse and worse. And she – I kept thinking that, well,
I can deal with this, so I took some kind of a pain pill and nothing helped.
(Charmaz, 2014, p. 119)
The codes show very clearly the general action orientation of grounded theory that
is particularly evident in Charmaz’ approach. Charmaz explicitly calls for codes to be
formulated preferably in the grammatical form of the gerund: ‘Experiencing escalating
pain’. This linguistic form of the codes expresses the fact that the focus is on action and
not on topics; however, the latter also occurs in Charmaz’s work, as the code ‘Inability
to control pain’, which is not in the gerund, shows.
The raw data come to life as researchers conceptualize about them. As Strauss and
Corbin noted, raw data are useless, for you cannot do much with them other than
merely count words or repeat what was said. Concepts are always named on the current
state of the analysis; they can also stem from the literature and do not have to be devel-
oped by the researchers. This can be advantageous because some such concepts are
associated with analytical meaning, such as ‘caregiver burnout’, ‘experience during ill-
ness’, and ‘loss of status’. However, many of these concepts are already connected to
certain theories, which could be disadvantageous to your current study.
During the first open coding of a text, we recommend that you pay attention to the
words and metaphors that the respondents use. In grounded theory, such words and
statements are referred to as ‘in vivo codes’. For example, Strauss mentioned the term
‘tradition bearer of the unit’, which the head nurse used to refer to another nurse in her
unit (Strauss & Corbin, 1998, p. 116; Kuckartz, 2010a, p. 75). In the remaining steps of
the analysis process, the grounded theory moves from the initial concepts to categories,
which are more abstract concepts or summarized concepts at a higher level of abstrac-
tion (Strauss & Corbin, 1998). For example, concepts such as ‘holding on’, ‘hiding’, and
‘moving out of the way’ may arise when watching children play. Those concepts could
then develop into the more abstract category ‘anti-sharing strategies’.
Concepts within grounded theory should be as precise and specific as possible. They
are not paraphrases, but they move towards a more abstract, general level. The follow-
ing examples of concepts stem from Strauss and Corbin (1996, pp. 106–107): ‘work in
the kitchen’, ‘attention’, ‘transfer of information’, ‘unobtrusiveness’, ‘timing of service’,
and ‘satisfaction of guests’. Once researchers have collected a good number of concepts,
they can group them together and summarize them.
Grounded theory can be seen as an invitation to and instructions for arriving at a
theory that is based on the data. The same idea of empirically grounded analysis is true
for QCA: every category and sub-category, every relationship presented, every evalua-
tion conducted, every typology constructed is rooted in the data, documented and
traceable for everyone who receives the research, whether they are experts or people
who read the research report. From the beginning, grounded theory aims to develop
theoretical categories (see Kelle, 2007); however, this is not necessarily the case in quali-
tative text analysis. Grounded theory does not require all of the data to be coded
because it focuses on moving forward, working with the categories, and developing
theories, leaving the data behind.
As a summary, the following points can be noted for inductive category development
within the framework of grounded theory:
• The open approach, in which everything that ‘comes to mind’ for researchers,
what is ‘induced’ by the text, is recorded in the form of codes
• The multi-stage process of category building, which seeks to achieve a higher
level of abstraction and analysis through integration and aggregation of those
earlier codes
• The focus on a few categories that seem particularly important in the course of
the analysis process
• Attention to words, terms, and metaphors used by the research participants
• Reflecting on the categories throughout the entire process of analysis.
are most suitable for your own study. Mixing different types of categories is also
possible. Furthermore, it is important to ask: How close do I want to stay to the
language of the research participants when creating categories? How far do I want
to work with more abstract categories or with analytical or theoretical categories?
For example, a respondent describes her activities in terms of waste separation. A
‘waste separation’ category can now be defined, but also more abstract categories
are possible, such as ‘individual behaviour in the area of recycling’ or even more
generally ‘individual environmental behaviour’.
3 Familiarize yourself with the data and determine the scope of the segments
to be coded. With reference to the remarks on hermeneutics in Section 1.4, it is
always important initially to familiarize oneself with the data and not to start
forming categories with the first line one reads. Those who have collected the data
themselves already have this familiarity, of course. In the case of very extensive
material, it may be sufficient to make a conscious selection, for example, to start
with particularly diverse interviews and then add others until a good overview of
the material emerges. Before starting to code and create categories, one should
also think about segmentation, that is, the scope of the text passages to be coded.
In the simplest case, only the term or the core of the statement leading to the
coding is coded. Alternatively, a formal criterion – a sentence, a paragraph – can
be defined as the segment to be coded. In most cases, especially when working
with QDA software, it is recommended that you code complete statements (‘units
of meaning’) that can still be understood out of context during later analysis.
4 Read the text passages sequentially line by line and develop categories
directly using the text. Assign existing categories or create new ones. The
development of categories can be done in a single-phase run through all the
data or through just a subset of the data. The order in which the texts are coded
is not really important, but the risk of bias should be taken into account and
randomly ordering the texts may help avoid this. The texts are now processed
line by line. For example, highlight and mark the text on paper and write
comments in the margins or use QDA software to do so electronically by
assigning codes to the given text passage. One should be aware that the
development of categories is a constructive activity on the part of the researcher
and therefore one should reflect on the nature of the categories that are
‘induced’ by the text. Do the categories tend to denote themes or are they more
analytical? How do I deal with particularly significant words or terms that
research participants use? Do I code these as natural categories (in vivo codes)?
The categories themselves may be comprised of single or combined terms or, as
in the case of argumentation and discourse analyses, of an argument, phrase, or
even a short sentence. The beginning of category development should be rather
open; no specific degree of concreteness or abstractness of the categories should
be prescribed yet. In the further process of category formation, the now steadily
complete the coding guide (cf. Section 2.5). The definitions should be illustrated
with concrete text passages (quotations). Special attention should be given to
the selection of these examples. They should be as typical as possible, clarify the
distinction from other categories and thus create confidence in dealing with the
category. How much data should you process in this way to arrive at an optimal
coding scheme? There is no fixed answer to this question, as you will have to
work through as many text passages as necessary until you have the impression
that no new aspects are surfacing within the text. Depending on the size and
complexity of the data to be analysed, this may be the case after you have
processed just 10% of the text; however, you may have to process as much as
50% of the text. Normally, you will have to process the data in several cycles:
some categories can be grouped together according to their similarity, while
other categories that may have started out too broad may have to be subdivided.
2
The study was conducted by the federal government in 2015 in the context of its
citizens’ dialogue on the understanding of quality of life. More detailed information is
documented and generally accessible on the internet at www.gut-leben-in-deutschland.de
(as of 20 March 2022).
1 Determine the goal of category formation based on the research question. The
goal of the study is initially explorative and descriptive: we want to find out something
about what Germans currently consider to be personally important in their lives. In
doing so, we want to find out which topics are in the focus of attention and which
terms are mentioned by the respondents. The frequency with which topics are
mentioned is also of interest. As a result of the research a report is to be written in which
the question of what is personally important to Germans should take up about 20 pages.
In the first step of category development, the main categories are to be formed. The
decision whether sub-categories should also be developed and if so, whether for all or
only for selected main categories, should only be made after a first coding cycle.
3 Familiarize yourself with the data and determine the scope of the segments to
be coded. Now it is important to get an overview of the data: many statements must
be read through until the impression of a good overview is achieved. Even in real life, it
is difficult to give exact information about how long it takes to get an overview. This is
of course strongly dependent on the ‘terrain’ over which one would like to get an
overview. In the case of this example study, the answers of the respondents are quite
short and so interesting that one keeps on reading spontaneously. A reasonable amount
of reading time should be planned: at least 100–200 answers to the question of what is
important to you personally should be read before you start to develop categories.
After a first overview of the data has been gained, two questions need to be
clarified next: the question of the scope of a segment to be coded in each case and
the design of the category system. The reading of the responses shows that not
only the scope but also the type of responses varies greatly: on the one hand quite
extensive prose texts, on the other hand only a series of keywords. Coding units of
meaning, the ‘gold standard’ of QCA, is therefore not always possible. It makes
sense to establish the following rule for this data. For responses formulated in
sentence form, meaning units of at least one complete sentence are coded. In the
case of keyword-type responses, only the words of the individual key points are
coded and, if necessary, only individual words.
With regard to the thematic categories to be formed, there is still the question
of how to deal with valences, that is, positive or negative valuations in the
answers given by the respondents. Take a statement like this: ‘A good life means
living in an environment worth living in. Large wind turbines that come
extremely close to our houses make a good life impossible.’ Here the topic of
renewable energy is indirectly addressed, but the topic has a negative
connotation. Is the conclusion that a decision is required whether each valence
expressed is positive or negative? The reading of the responses shows that
negative evaluations only occur very rarely, which is understandable in view
of the positively worded question ‘What is important to you personally in life?’
Therefore, in this case it is appropriate that no specific categories are created for
the valence, but it is taken into account in the later analysis of the category.
There is one more point to clarify, which has to do with the nature of the
questions asked in this online survey. Respondents had been asked to give their
statement a heading. How do you deal with the headings when coding the
data? Should the headings be treated like the other text of the response and
coded in the same way? A specific check showed that the headings often not
only sum up the text but also contain information of their own. In this respect,
it is logical to treat the headings like normal text when creating categories. At
the same time, it can be assumed that headlines, similar to those in daily
newspapers, still have a special informative value. When working with QDA
software, headlines should therefore be marked and coded as such (with the
code ‘Headline’) and then subjected to a separate analysis.
4 Read the text passages sequentially line by line and develop categories
directly using the text. Assign existing categories or create new ones. Starting with
the first text, proceed line by line. Figure 3.5 shows the answer of one person in
the left-hand column and the codes developed, some of which already represent
potential thematic categories, in the right-hand column.
If the aspects that appear in the answers are openly coded, as has been done
here, one should proceed quickly and not think too long about the best choice
of words when formulating codes. In the next step of the analysis, all codes are
ordered and systematized – there is still time enough to find the most
appropriate and fitting words for the categories. This is – by the way – also a
great advantage compared to the development of categories via paraphrasing
and focused summarizing, where one tends to spend a lot of time on
formulating paraphrases. Also, no premature generalizations or abstractions
should be made and far-reaching interpretations should be avoided.
Response Codes/categories
Time for family and friends, mental and physical Time for social life
FREEDOM, nature Mental and physical freedom
Friends: this requires like-minded people and time. Many Family
people around me have little time to meet with each other. Friends and like-minded people
Instead: Everyday stress (lots of work, lots of driving the
kids around), short messages via WhatsApp. For me and
my children, I would wish for more people with time and Leisure without stress
less media communication and planned leisure time – in
other words, children who play outside (also in the forest!)
with lots of time, as they used to!
Freedom: The woman in the burqa in the supermarket Children’s play in nature
the other day threw me off. Many Muslim women wear
headscarves and cover their bodies with long clothes.
Muslim women did not shake hands with my husband at Cultural commonality
the barbecue recently. Why, I asked – ‘Because it says so
in the Koran’. For me, these women and especially girls are
restricted in their physical and mental freedom (criticism Personal freedom
of religion). I think: Religion must not take precedence Self-determination
over the ideas of the Enlightenment (Kant) and the Ideas of the Enlightenment
emancipation of women. Emancipation of women
Response Codes/Categories
All five are Important
I need Family
a family, today often a patchwork family, including the children – Children
who live with my ‘ex’ (but which is severely restricted for me). material security
a job that feeds me and is fun as often as possible, including fun work
employee rights that (somewhat) protect me. Security through workers’ rights
Friends who support me and that can also be people from other
cultures and social classes. Friends
Diversity
Health, healthy food, good medical care that is not solely Health
oriented towards the monetary interests of big business. Supply of healthy food
Medical care
Safety, I want to be able to move around without fear, even at Safety in the city
night in the metro
5 Systematize and organize the category system. Near the end, when it
becomes difficult to keep track of all the codes and categories that have been
formed, the work process should be interrupted and attention focused on the entire
system of categories. The full set of codes and categories, at whatever level of
abstraction, that have been created so far are now ordered and systematized. How
does this work? First, identical or very similar codes can be merged into one code.
Second, codes can be bundled together into a new or already existing category.
Third, the overall shape of the category system should be thought about. Do the
category system and the individual categories support the answering of the research
questions? Is the category system coherent in itself? Is the relation of the categories
to each other plausible? Is the category system exhaustive?
It also makes sense to start writing category definitions, which can be illustrated
immediately with quotations from the already coded data.
Organizing and systematizing the codes and categories is technically best done
by writing them down on moderation cards and using a large working surface
(pinboard, tabletop, etc.) or – much more convenient and effective – with the help
of QDA software. Here, the codes and categories developed can easily be organized
and grouped on a sufficiently large-screen work surface. This way of working also
has the advantage that there is always a link between the categories and the
original text.
Figure 3.7 shows a category system3 consisting of eight main categories, an
intermediate result of the process of category development. The main categories are
arranged around the centre of an imaginary individual in this graphic.
The illustration was created with the function ‘Creative Coding’ in MAXQDA.
3
Creative Coding
Promoting talent
Education Kids
Like-minded
Freedom
people
Secondary network
Job satisfaction
Social welfare Friends
Health
Social interactions
Time Cultural homogeneity
Figure 3.7 Systematizing and ordering codes into categories with QDA software (reprinted
by permission from: Springer Nature. Analyzing Qualitative Data with MAXQDA by Udo
Kuckartz and Stefan Rädiker © 2019)
The codes, so far developed in the process of open coding, are connected to the
main categories by directional arrows. Behind the codes in the outer circle of the
diagram are concrete text passages – at least one text segment. The codes are
illustrative examples of what is hidden behind the main categories; in this case,
they are not sub-categories, because the development of sub-categories was
deliberately dispensed with at this point in the analysis.
As soon as a satisfactory order and systematization has been achieved, the
coding process continues with the processing of further data. Of course, the results
of the systematization step are taken into account when assigning further text
passages to already existing categories, that is, the new main categories are used
and, if necessary, supplemented by new codes and categories.
6 Finalize the category system and create category definitions. If nothing new
or only singular aspects (i.e., only aspects specific to the life situation of a particular
person) appear, the process of category development is terminated. Now follows the
step that is significant for the coding of the further material, namely the final
specification of the category system. As stated in the guidelines above, this does not
mean that no more changes are permitted to the category system in the future. The
final specification of the category system also includes formulating category
definitions. Since the codes grouped into a category are assigned to existing text
passages, it is easy to include concrete examples from the data in the definitions.
When working in a team, it makes sense to do the step of specifying the final
category system together. It is important now to have the research question(s) in
mind and to ask ourselves: Where do we want to go with the final category scheme
and the subsequent coding and analysis using the categories? What do we want the
product of our analysis to look like? In this example, the goal of the content
analysis is to identify the topics that the research participants consider important
for their lives and then, in the second step of the analysis, to get an overview of
what is hidden in detail behind the major topics; in other words, to find out what
exactly is declared to be important when, for example, someone talks about nature
being very important to them personally.
There are now several strategies that can be adopted with regard to the final
category system. One variant involves the development of a relatively large number
of quite specific categories (20 or more). This strategy, which is often adopted in
quantitative content analysis, is not adequate for qualitative research. It is more
appropriate to keep the number of main categories smaller and above all to pay
attention to the relation of the categories to each other – the shape (gestalt) of the
category system.
The development of the category system is not just preparatory work for the
analysis which will follow, but already a part of it and represents an analytical
achievement, which should also be presented in the research report in appropriate
detail. The report does not have to be limited to describing what is present but can
also point out what is missing. In our example, in the selection of statements made for
the development of categories, it is striking that an intact close environment (partner,
children, family, friends) is very often considered important, as is the possibility of
social influence in the sense of democratic participation, but that the middle ground
of an intact neighbourhood or intact community is only mentioned relatively rarely.
(2005), Gläser and Laudel’s (2010) approach, as well as the methodically well-
documented study on family and right-wing extremism by Hopf et al. (1995). In the
latter, hypotheses and categories were formed first, derived from attachment theory.
These categories were then assigned to the data, specified, modified, and differentiated
as necessary. At the same time, unexpected elements in the data, such as those that
could not be derived from the attachment theories, served as inspiration for new cat-
egories (Schmidt, 2010). The mixture of deductive a priori category and inductive
category development based on the data occurs almost exclusively in one way: deduc-
tive categories are developed first and the development of categories or sub-categories
using the material follows in the second step, which is why one can also speak of
deductive-inductive category development.
Depending on the research question and schedule of a given project, there are differ-
ent ways to develop categories in a deductive-inductive fashion. The general process is,
however, always the same. You start with a category system that contains relatively few
main categories (usually not more than 20) that have been derived from the research
question, the interview guide and/or a reference theory. Unlike within pure deductive
approaches, these categories are seen only as starting points here. They serve as search
aids, meaning that you can search through the data for relevant content and roughly
categorize it. In the second step, sub-categories are developed inductively, whereby only
the material assigned to each main category is used.
A typical application of deductive-inductive category development is described in
Chapter 5 on structuring QCA. In Rädiker and Kuckartz (2020) we present a method
that works with such a deductive-inductive approach, in which the so-called basic
categories are primarily developed from the interview guide and further differentiated
by sub-categories developed using the empirical data.
deductive or inductive category formation in its pure form, but in most projects mixed
forms are practised, often in such a way that relatively broad categories are initially
formed, which are then gradually elaborated on in a data-based manner in the further
analysis process. This strategy of deductive-inductive development of categories offers
the advantage that it can be linked initially to the current state of research, which also
strengthens the feeling of being on safe ground and provides legitimacy for one’s own
research work.
When creating categories, it is important to keep the whole in mind – the entire
system of categories. This analytical framework should be coherent and have a consist-
ent and plausible overall shape so that it can be easily communicated in the scientific
community. It must be clear at a glance, so to speak. In concrete terms, this means that
the categories should be set at a similar level of abstraction and the category names
should be formulated in a consistent way.
Chapter objectives
In this chapter you will learn about:
Within social research there are a number of methods and techniques for qualitative con-
tent analysis. Mayring (2021), for instance, distinguishes between nine different forms
of QCA: (1) summarizing, (2) inductive category formation, (3) narrow and (4) broad
context analysis, (5) nominal deductive category assignment, (6) ordinal deductive cate-
gory assignment, (7) content-structuring/theme analysis, (8) type analysis, and (9) parallel
forms. However, some of these techniques, such as content structuring and type analysis,
are only very briefly outlined there.
In the following chapters, three basic methods of qualitative content analysis are
described in detail:
• structuring QCA, which can be regarded as the core method of QCA procedures
and in which the material is typically coded in several coding cycles with
deductively and/or inductively formed categories (Chapter 5)
• evaluative QCA, in which the material is assessed by the coders and assigned
evaluative categories (Chapter 6)
• type-building QCA, which is usually based on structuring and/or evaluative
analysis and whose primary objective is to develop a typology (Chapter 7).
These three methods use different strategies of analysis and are each used very fre-
quently in research practice. This is especially true for the first type, structuring QCA.
Interestingly, structuring analysis in form of thematic analysis (Früh, 2017) is also by
far the most commonly used method in the field of quantitative content analysis. In
this area, it is mostly done as frequency analysis of themes. What distinguishes qualita-
tive content analysis from quantitative analysis can be seen particularly well in the way
the themes are analysed. While the atomizing manner of quantitative analysis aims to
convert the verbal data into precise categories (represented by numbers) and then to
statistically evaluate the resulting data matrix, qualitative content analysis is interested
in the text itself, notably based on the text in its entirety. Even after categories have been
assigned, the text itself (i.e., the wording of the statements) is relevant and also plays an
important role in the preparation and presentation of results. In quantitative content
analysis, however, the results include merely statistical parameters, it is coefficients and
models that are interpreted and presented. After the coding process, the verbal data
in quantitative content analysis are no longer of interest, even as quotations, because
the plausibility of the results of the statistical analysis cannot be demonstrated using
selected text passages.
All three QCA methods described below include both topic-oriented and case-
oriented methods. This means they can be viewed not only as category-based analysis
but also at the case level as case analysis (e.g., in the form of case summaries). The
comparison of cases or groups and clusters of cases plays an important role in the
analysis process for each of the three methods. This also marks an important difference
from Mayring’s conception of QCA: there, the case-oriented perspective plays almost no
role compared to the category-oriented perspective (Steigleder, 2008, p. 174).
Present
results
Read and
explore
Analyse Research the data
coded data questions
Develop
Code categories
data
The sequence research question → select or collect data → data analysis is quite charac-
teristic of all forms of empirical research, and QCA is no exception. In Figure 4.1, the five
phases of QCA are illustrated, starting with the research questions placed in the centre,
surrounded by the subsequent steps. On the one hand the figure shows the basic
sequence, but on the other hand it also illustrates the possibilities for circular processes
mediated by the research question. The core of every QCA is represented by the phases
‘develop categories’, ‘code data’, and ‘analyse coded data’, whereby the forming of cat-
egories and the coding of data can take place in several cycles. The figure is meant to
illustrate that the flow of the analysis process is much less linear than in the classical
model of the research process and that the different phases of analysis are not strictly
separated from each other. It is even possible to acquire additional data after the category
system has been established and the majority of the data has been coded. The figure also
illustrates that the research questions play a different role in QCA than in other
approaches. They are posed at the beginning of the research process, but they do not
remain unchanged as in the classical hypothetical-deductive model only to be answered
at the end of the analysis. Instead, the research question is central to each of the five
method areas and can be changed dynamically during the analysis process (within cer-
tain constraints). For example, you may wish to make the research questions more
precise, place new aspects at the forefront, or modify the research questions because of
unexpected discoveries.
The phases centred on developing and modifying categories in which researchers
progressively work with the data are very important phases of the analysis process. Even
if a QCA is conducted in a theory-based manner with hypotheses in mind – which is not
out of the question – you can fine-tune the categories during the analysis process and
add new (sub-)categories if you deem it necessary while working through the data.
What role does theory play in QCA overall? Based on the identification of QCA as a
method, the question needs to be asked differently: what role does theory play or
should theory play in a research project in which QCA is used as a method of data
analysis? That is, the role of theory in the specific research project must be questioned.
There may be projects that want to describe a certain social phenomenon as precisely
as possible (e.g., ‘How do young fathers experience paternity leave?’) and do not aim
at theory building at all. Other projects are theory-oriented and take a specific theo-
retical approach such as the theory of planned behaviour (Kan & Fabrigar, 2017) as
their starting point, for example, to investigate the connections between environmen-
tal attitudes and behaviour. Yet other projects explore a topic in a very open manner
but aim to generate theory through their research. In all these projects, QCA can be
used beneficially to some extent independently of the role of theory. In the process
model outlined above, theory can serve both as a starting point in determining the
formulation of the research questions and as the end point of a very open approach.
Thus, the objective can be the generation of theory, the confirmation of theory, the
refinement of theory, and finally, as mentioned, projects may also want to deliberately
focus on description. Intentionally, we do not speak of ‘limiting’ here, because this
sounds as if such research projects have a deficiency and, in a sense, stop halfway.
However, the usefulness of a project (e.g., for addressing social problems or with regard
to transformative questions) does not depend on whether theory building is the pri-
mary objective of the project. Rust (2019) suggests that theory generation should be an
explicit phase in the process model of QCA. However, the different roles of theory in
research projects argue against the integration of such a phase in the general process of
QCA. On closer inspection, it becomes clear that Rust’s proposal, based on experiences
in teaching research projects, also refers only to the second type of project (open
approach with the goal of theory building at the end). However, Rust points out a very
significant problem, namely that in many projects the role of theory or of theory build-
ing is not sufficiently considered and that before writing down the results, a review of
the state of research and the literature should be carried out.
Intensely explorative or descriptive research will perhaps focus on the analysis of issues
and arguments, examining the relationship between categories or, in the style of grounded
theory, work to create core categories for the phenomena identified in the research field
(Corbin & Strauss, 2015, pp. 187–190). In these cases, neither evaluative nor type-building
analysis would be appropriate. Evaluative analysis would force judgement too early in the
process, and both methods follow approaches other than the comparative method of
grounded theory, which operates mainly in terms of minimum or maximum contrasts.
In the literature on QCA, there is agreement that structuring content analysis is the
core content analysis procedure (Mayring, 2015; Schreier, 2012, 2014; Steigleder, 2008).
For this reason, it takes up most of the space in this book. In a contribution comparing
the different variants, Schreier (2014) notes that evaluative and type-building QCA are
also structuring content-analytical procedures at their core. This is certainly true, but
the processes of evaluative and type-building analyses are so different that separate
descriptions seem appropriate – and of course the methods can also be combined.
What do the three methods of QCA presented in Chapters 5–7 have in common? We
can identify six key points:
1 They are methods of analysis, that is, they do not prescribe a specific type of data
collection. Different methods, such as structuring and type-building QCA, can
easily be applied to the same data, for instance, during the secondary analysis of
existing qualitative data.1
2 They are methods that compress and summarize the data rather than expand in
the style of sequential analysis in order to interpret the data exegetically.
3 They are category-based methods. Thus, the analytical categories are the focus of
the analysis process, although the way in which the categories are constructed may
vary. Categories or topics can be developed from the theory or the research
questions and applied to the data, or developed directly grounded on the data. It is
quite common to combine various methods to build categories.
4 They are systematic scientific methods and not open to artistic interpretation,
meaning the implementation of these methods can be described precisely and
mastered by researchers and students. These methods do not involve the art of
interpretation that is characteristic in fields such as literary or art history.
5 The three methods are language-related and are initially conceived as methods for
systematic qualitative analysis of verbal or textual data. Nevertheless, they may also be
applied to images, movies, and other products of culture and communications.
1
For more information on secondary analysis of qualitative data, see Medjedović and
Witzel (2010) and Medjedović (2014). Qualitative data for secondary analyses are provided
by Qualiservice, a research data centre at the University of Bremen (www.qualiservice.org),
and by the UK Data Service (https://discover.ukdataservice.ac.uk/qualibank).
For all three methods, analysis can begin even before all of the data are collected. All
three methods are compatible with different sampling methods and can be combined
with more conventional methods, such as a quota-based sampling, or theoretical sam-
pling as is preferred in grounded theory. However, as systematic processes, the three
methods require a complete coding of the entire data material, meaning that later
changes to the category system require additional processing of the data and are there-
fore associated with substantial additional effort. The postulate that the entire data of
a study can be analysed in a systematic, category-based manner prevents researchers
from drawing premature conclusions based on individual cases. Regardless of which
analysis method you use, it is recommended that you document the individual steps of
the analysis process as accurately as possible in a research journal.
• Our form of QCA always begins with an introductory phase that is primarily case-
oriented. The deliberate entry into the analysis is intended, among other things,
to help researchers not to equate QCA with only coding and categorizing text
segments and not to lose sight of the overall picture by hastily starting to code.
• Regarding the interpretation of texts, we are guided by hermeneutic principles as
the classical approach to understanding texts.
• Especially in the introductory phase, but also later, word-based analysis methods
such as keyword-in-context or word frequencies can be integrated.
• The procedure for all three types is divided into phases and described in as much
detail and as concretely as possible.
• The procedure considers the support of the analysis by QDA software from the
outset. Accordingly, the book contains a detailed chapter on the implementation
of the procedures using computer software.
• One focus of the descriptions is on the phase after coding – the analysis of the
coded data – for which numerous procedures and possibilities beyond mere
category frequencies are presented. In this context, we also show possibilities for
summarizing coded segments and integrating this into further analysis.
• A frequency analysis of the categories in some situations makes sense (if the
number of cases is large enough), but the actual analysis is qualitative.
• Despite the focus on the research questions and the postulate of methodological
rigour, there is a place for serendipity (Merton & Barber, 2004), the accidental
discovery of important findings that were not the primary focus of the research,
but which prove to be significant for the object of investigation.
• Continuous writing of memos to record working hypotheses, analysis ideas, and
findings is a recommended part of the whole analysis process.
• No pre-definition of segments to be coded is required as the first step of the analysis.
We recommend this only for the calculation of chance-corrected coefficients of
intercoder agreement.
• Consensual and discursive verification of intercoder agreement is prioritized over the
mere use of coefficients.
• Theory orientation is not per se the starting point for the analysis, analytical
decisions, or the selection criterion for inductive category development, although
the inclusion of theories can be valuable at many points in the analysis process.
Theory orientation is also not regarded as the primary quality criterion by which
a QCA should be evaluated.
• No predetermination of the level of abstraction of categories is postulated. But, of
course, we consider a similar level of abstraction across the main categories as an
important criterion for the assessment of the quality of category systems.
• During the entire analysis process, the focus can be directed both to the case
level and the category level. Depending on the research questions, the case can
remain relevant in its entirety, including its multi-perspectivity, its
multidimensional nature, and its possible inherent contradictions. The case does
not necessarily lose its importance because of cross-case coding. Case orientation
and category orientation can also be combined in a study. This ‘double perspective’
of cases and categories marks an important difference from Mayring’s conception
of QCA: there, the case-oriented perspective plays virtually no role compared to
the category-oriented perspective, as the following quote illustrates:
Text analysis is thus selectively limited to the category system. Contents that
are not addressed in categories, as well as a holistic impression of the text, are
not taken into account or would have to be addressed with other text
interpretation procedures. (Mayring, 2019, para. 3)
The list could be continued in detail, but this would not be very useful. We have already
mentioned further aspects in Section 1.9 when we pointed out the central characteris-
tics and focal points of our definition of QCA in contrast to other definitions. Further
explanations of the similarities and differences between various approaches to QCA can
be found in Schreier (2014).
• In a grounded theory study, the goal is always to develop a new theory (of
medium scope) grounded in the data. This can also be a goal when using QCA,
but it does not have to be.
• To achieve the purpose of theory building, it is not necessary to code all the
material in a grounded theory project, because the end point of the analysis is the
saturation of the theory generated, not the coding and category-based analysis of
the complete material. It is also not necessary in grounded theory approaches to
process and code all the material in the same systematic manner as is done in QCA.
• In a grounded theory study, changes to the category system are possible on
an ongoing basis; in fact, continuous work on the codes and categories is a
characteristic feature.
• The grounded theory approach not only describes an analysis method, but also
includes data collection (its keyword is ‘theoretical sampling’). Optimally, data are
continuously collected and analysed; overall, the approach is very process-oriented
and flexible. In contrast, projects that use QCA as a method of analysis usually
work with predetermined samples. Such systematic selection plans make it
possible, among other things, to conduct comparisons of sufficiently large groups.
• Due to its objective, grounded theory approaches are less suitable for descriptive
analyses, and the same applies to their use in evaluation.
At this point we would like to point out a misunderstanding that we have often encoun-
tered in practice. Many times we have been presented with projects labelled as grounded
theory studies which, on closer inspection, turned out to be qualitative content analyses
with inductive category development. The mere fact that categories are developed using
the data in the style of open coding does not justify a designation as a grounded theory
study if there is no intention to work conceptually and analytically towards a theory (a
keyword here is ‘theoretical sensitivity’).
Yet, as I showed in the last chapter, numbers have a place within qualitative
research, assisting, for example, in sensitive attempts to learn lessons in one
place that have relevance for actions and understanding in another place.
There is a variety of other uses of numbers which can enhance the quality of
qualitative research. (Seale, 1999a, p. 120)
As a result of his very instructive overview of the benefits and use of numbers in qualita-
tive research, Seale (1999a, p. 121) formulated the principle of ‘counting the countable’.
Numbers can assume different functions; they can not only be simple frequencies or
percentages but also be used for more complex statistical calculations, such as crosstabs
with the chi-squared test or cluster analysis. They can clarify arguments and support the-
ories and generalizations. Seale’s (1999a, p. 138) emphasis on ‘avoiding anecdotalism’
expresses the importance of using numbers quite concisely.
However, there should be careful reflection on the significance that quantification
can have for every QCA. The main question here is: does frequent also mean important?
Counting what can be counted always means that one acquires additional information
but does not relieve one of the task of reflecting on the significance of quantification
and statistics for the particular research project. For example, if in an online interview
with relatively short answers a question asks about what is personally important in
one’s life, it is justified to describe in the analysis a topic mentioned by more people as
‘important to more people’ than a topic mentioned less frequently. So, if the category
‘primary network’ was coded for more people than ‘secondary network’, it is legitimate
to write in the research report that the primary network is considered important by
more research participants than the secondary network. The situation is different at the
individual level: if someone has mentioned ‘primary network’ more often in relatively
short answer texts, it is difficult to justify the conclusion about relevance.
One should also be careful not to draw conclusions from the category-based fre-
quency analysis of the data about questions that were not asked at all. Assuming that,
in the analysis of the question about what is important in one’s life, the category ‘secu-
rity’ was developed with corresponding sub-categories ‘internal security in public space’,
‘external security, peace’, etc., it cannot be concluded from a higher frequency of ‘inter-
nal security in public space’ compared to ‘external security, peace’ that internal security
is more important to the research participants than external security. This was not the
question that was asked. In general, it can be said that the further one moves away from
the questions asked, for example in a guided interview, the less meaning numbers have.
They remain important as additional information, but the meaning must be critically
considered in the specific case. The more openly the interview was conducted, the less
useful frequencies become.
A conscious approach to quantification in QCA requires reflection on what should
actually be counted or what can be counted at all given the present data and the coding
used. In this respect, it is helpful to take a look at quantitative content analysis, which
uses the term unit of analysis (also called recording unit). This unit of analysis indicates
which units are processed and counted. When analysing newspaper reports in quantita-
tive content analysis, for example, the entire newspaper report, a paragraph, sentence, or
even an individual word can be selected as a unit, which must then be coded accordingly.
Meaning units can also be chosen as the unit of analysis, but they must be defined
beforehand and the text must be segmented accordingly. In one and the same study,
analyses can be carried out with different units of analysis, that is, the definition of the
unit of analysis and the segments to be coded should be well thought out at the begin-
ning of the content analysis, but the units of analysis can vary during the analysis.
In a QCA of newspaper reports, the analysis can also be carried out at the levels just
mentioned, and the same applies to the analysis of interviews. Mostly, however, one will
not adopt formal units of analysis here, but work with units of meaning, which do not
have to be defined in advance. If coded segments of a category (i.e., the category frequen-
cies) are counted, this can be done per interview or also per group of several interviews
if these groups are to be compared with each other. It should be noted, however, that
high category frequencies in one interview do not necessarily imply a high importance
of the category for this interviewee. If the category ‘learn dealing with global problems’
was coded many times in an interview, this does not necessarily mean that the person
mentioned many different ways of learning, but he or she may have made the same sug-
gestion several times in short statements. Frequent, therefore, does not necessarily mean
diverse and relevant. Consequently, group comparisons are usually based on the number
of cases in which a category is coded, rather than the number of coded segments. Instead
of asking how many times a particular category was assigned per case, one should ask
whether the category was assigned to the particular case or not, or to how many cases
the category was assigned. A special situation arises in the case of focus groups and group
interviews, because here either the entire interview or the individual person can be
defined as the unit of analysis, so in a sense another hierarchical level is added.
What else should be considered when quantifying in qualitative content analyses?
The primary goal of working with quantifications does not have to be the presentation
of frequencies, percentages, and statistics in the research report. Quantitative analyses
can also serve as an exploratory tool and provide entry points into more in-depth analy-
ses. Be careful when working with percentages because the presentation of percentages
is not useful when the sample is very small. For example, reporting case percentages with
five interviewees is considered to be a methodological mistake, because here one case is
already equal to 20%. Normally, the calculation of percentages with a sample of less than
20 cases is not very meaningful. Also beware of supposedly increased generalizability
through quantification: the inclusion of statistics in the analysis and the presentation of
results does not per se guarantee generalizability, because this is primarily a question of
sampling and the type of conclusion (cf. Section 9.5).
like to examine? What sorts of preliminary assumptions do I have regarding these rela-
tionships? For whom or what do the results have a use? Likewise, it is helpful to reflect
on your own prior knowledge and assumptions about the object or phenomenon under
investigation – preferably in writing (Rädiker & Kuckartz, 2020, pp. 26–28).
Clarifying your objectives in this manner does not violate the principle of openness,
which is frequently named as a characteristic of qualitative research. The postulate of
openness refers first to the process of data acquisition, as respondents should have the
opportunity to express their own views, use their own words instead of being forced to
use predetermined categories, and express their individual motives and reasoning. On
the part of the researcher, openness in the sense of the principle of ‘approaching the
project without any research question and without any concept’ would not only be
mere fiction (because we always operate on the basis of prior knowledge and prejudices
and a knowledge of the world that precedes any observation), but also imply ignorance
of the scientific community in which one moves. After all, in most situations there is
already a long-standing tradition of engagement with the object of research. However,
openness on the part of researchers is very much necessary with regard to openness to
other perspectives and interpretations as well as openness in the sense of reflecting on
one’s own prior knowledge and existing ‘pre-judgements’.
We cannot analyse our data unless we read it. How well we read it may
determine how well we analyse it. … The aim of reading through our data is to
prepare the ground for the analysis. (Dey, 1993, p. 83)
For systematic data exploration taking hermeneutic principles into account, it is helpful
to first visualize the context in which the data were generated. In an interview study, one
can ask oneself questions such as: What information did interviewees receive in advance?
In which settings did the interviews take place? What expectations might the research
participants have had of the interview? When starting the analysis, it may also be possible
to refer back to the actual raw data in the form of audio and video recordings. This begin-
ning of the first phase of QCA is called initial work with the text, whereby – as in literary
studies – working through the text is to be understood as intensive exploration of the
content and the language of a given text. Starting with the first line, you must read the
text sequentially and completely. The goal is to gain a general understanding of the given
text on the basis of the research question(s). It is often helpful to outline your research
questions and attempt to answer them as you work though the interview data. For exam-
ple, in a study about individual perceptions of climate change, we were able to answer or
address the following points while reading through each of the interviews:
It can also be useful to examine the text formally. How long is the text? Which words
are used (particularly noticeable words)? What sort of language does the respondent use?
How long are the sentences? What sorts of metaphors are used?
So, what does it mean to systematically read and work through a text? Reading is an
everyday skill that we have mastered, and we have developed a variety of individual read-
ing techniques within the sciences. Some people highlight texts with one or more
multi-coloured highlighters, some people write notes in the margins using their own
abbreviations, and others record their notes on other pieces of paper, index cards, or in a
research journal. The list of such individual techniques that have proved effective over
time is long; such methods should by no means be considered inappropriate here.
However, there is a strict procedure that must be followed in QCA in order to ensure its
comparability, understandability, and methodological consistency, and it will be explained
in more detail in the steps below. Moreover, it should also be noted that many of the tried
and tested methods mentioned above (highlighting, etc.) can also be used in QDA soft-
ware programs (cf. Chapter 8). In addition to electronic text markers, word-based analysis
options are also available which support the initiating text work in a meaningful way and
which are almost impossible to use without software. With QDA software, frequently used
words and word combinations can be identified, and using a keyword-in-context display,
the question of within which formulations selected terms are used can be answered.
Results of such analyses can provide initial answers to the research questions or offer
valuable suggestions for how to proceed with the analysis. In an analysis of the first 50
transcripts from the multi-award-winning podcast on the Covid pandemic with virologist
Christian Drosten (in German), we became aware of the two words ‘natürlich’ (‘of course’)
and ‘vielleicht’ (‘probably’). Both occur about 650 times and are thus among the 15 most
frequent words, if you exclude definite and indefinite articles as well as other typical ‘stop
words’. In what contexts are these two words used? To what extent do they express a
degree of certainty in the statements made? Do they occur with specific topics? Are they
used more by the interviewer or by Dr Drosten? With QDA software, these questions can
be answered directly and the results recorded. Regardless of this, it can be stated for the
further analysis that the degree of certainty with which statements are made in the pod-
cast should be taken into account as an analytical perspective. It may be possible to
develop (evaluative) categories for this at a later stage.
Word-based methods can refer to several cases or be applied to individual texts.
Although the initial work with the text is primarily focused on the individual case, the
perspective on the entire data always remains important, especially in order to be able
to assess and situate the individual case in the overall context and to select specific texts
for comparison, for example a text in which the word ‘maybe’ occurs frequently.
When starting the analysis, you should include as much data as possible to get a
comprehensive picture and to create a solid foundation for the following analysis
phases. With very large amounts of data, a sample may be selected. Normally, a quota-
based selection is appropriate for this purpose, in that cases are selected from each group
under study. A random selection or a selection based on relevant attributes such as the
complexity and linguistic characteristics of cases is also possible. In our example of the
Covid podcasts, the date of broadcast also plays a role and it would make sense to select
podcasts from the beginning of the pandemic as well as at several later points in time.
Whatever the situation, the selection should be made in such a way that the widest
possible range of data is achieved with maximum variation.
In qualitative research, there is not a very strict distinction between the phase of data
collection and that of data analysis as there is in the classical model of quantitative
research. This is also true for QCA. In contrast to statistical analyses of standardized
data, you do not have to delay the analysis until all of the data has been collected. Thus,
normally, you can start analysing the data you already have while continuing to collect
additional data. Even if you do not adhere to the grounded theory, in which data acqui-
sition and analysis are explicitly crossed, it can be beneficial to start analysing the
content before all of the data has been collected. Hence, you can start reading and work-
ing through the first interview as soon as it has been transcribed.
Let us summarize once again:
• Determining the sequence for working with the texts and making a selection if necessary
• Analysing the text in light of the research questions
• Reading the text intently
• Highlighting central terms and concepts
• Marking and making note of important sections
• Marking passages that are difficult to understand
• Analysing arguments and lines of argumentation
• Examining the formal structure (length, etc.)
• Identifying the internal structure (paragraphs, breaks, etc.)
• Directing your attention to the general progression of the text
• Examining frequently used words and phrases and considering selected words in their
context.
The above list is not intended to imply that all exploration techniques must be used in
every analysis. For example, while it can be very relevant in an analysis of party manifes-
tos how long the sections on taxation are, the length of text passages in interviews may
be much less significant. Generally, it is advisable to go through the list of exploration
techniques systematically and consider which ones are used in one’s own project, with
what objective, for what gain in knowledge, and for what reasons other exploration
techniques are dispensed with.
After completing the initial work with the text, one should be familiar with the data
and especially with the individual cases. Furthermore, the results of this step can be very
diverse. Initial text work can provide useful insights for the design of the category sys-
tem, indications of the relevance of themes and aspects, as well as ideas for further
analysis. Furthermore, attention is drawn to important terms, metaphors, and formula-
tions and relevant, possibly particularly striking, text passages are identified. It is also
possible to identify salient cases, such as those in which interviewees hold extreme
opinions. All in all, the initial work with text brings the research questions closer to the
data and triggers insights into how to approach the material in order to answer the
research questions. Last but not least, this step of the analysis can also produce surprises
and serendipitous findings.
Writing Memos
Whether you work directly on the computer screen or rely on a printed version of the
text depends on your own personal work preferences and style. Many people find it
helpful to read through a printed version of the text first so that they can make notes
in the margins and highlight passages that seem particularly important. If you choose
to do so, make sure that the paragraphs or lines are numbered so that it is easier to
transfer your markings and comments into the electronic version later. On the screen,
you can highlight important or notable passages using an electronic highlighter.
Any peculiarities in the text or ideas that you may have while reading the text should
be recorded as memos.
When writing memos, it is helpful to be
aware of what kind of content is being noted
A memo contains any thoughts, down. A distinction can be made between
ideas, assumptions, or hypotheses descriptive, interpretive, and analytical memos.
that occur to researchers during the Description primarily means the descriptive
analysis process. Memos can be summary of data, while interpretation goes
short notes (like a sticky note) or beyond this and includes the researcher’s inter-
more reflective comments regarding pretation (e.g., ‘cause could be’). Memos
the content which act as building regarding the further analysis process are more
blocks for the research report. organizational in nature and may include deci-
Writing memos should be consid- sions and outstanding tasks (e.g., ‘Check in
ered an integral part of the research other cases too!’). To distinguish the different
process. types of content in a memo, it can be useful to
format the content differently, for example,
using a particular font colour.
Grounded theory addresses the role that memos play in the research process in detail
(Charmaz, 2014; Strauss & Corbin, 1990, pp. 197–223) and differentiates between
different types of memos. While memos are not quite as important within the framework
of QCA as they are in grounded theory, they are considered helpful resources that can be
used throughout the entire research process, just as they can be in grounded theory.
Case summaries can also take the form of a detailed, fluid text. They can be written
not only for individual interviews, but also for qualitative studies with groups and
organizations using mottos. A motto can focus on a particular aspect of the research
question, be based on a statement or quote in the given text, or be creatively formulated
by the researchers to fit a given text. However, because mottos are accentuated charac-
terizations, they are highly subject to interpretation. Thus, while a motto can be useful,
it may not always be.
The case summary in Figure 4.5 is headed with a concise characterization. It is taken from
a study of the graduates of the first two years of the Marburg BA programme in Educational
Science. It is an example of a more detailed, fluid text. It is written using the guideline struc-
ture and questions, which are highlighted and precede the respective paragraphs.
For interview studies, it is usually advisable to prepare case summaries, whereby a
well-reasoned sample should be selected in studies with larger numbers of cases. By
preparing case summaries, one gets an overview of the spectrum of cases included in the
research, which is of great value especially in studies with larger numbers of cases.
According to the criterion of maximum and minimum contrast, one can compare cases
that are particularly similar or particularly dissimilar to each other.
How can case summaries be handled for types of data other than interviews? This is
highly dependent on the characteristics of the material and the answer to the question
of what should be considered a case. If you are studying the role of the father in chil-
dren’s picture books, the case is the individual picture book and the production of case
summaries certainly makes sense. In focus groups, there are two possible definitions of
what a case is: either the respective focus group as a whole or an individual participant.
The more important differentiation by individual participants is for the analysis, the
more useful it is to write case summaries for individuals. If a differentiation by partici-
pants is not intended, case summaries for the respective focus group are more
constructive. In surveys with open-ended questions, on the other hand, where there are
usually short answers from a very large number of people, case summaries do not make
sense because summarizing the short answers would not bring any analytical benefit.
The same applies to manifestos of political parties, as these have already been written
with the aim of providing highly condensed and thematically structured statements.
Case summaries are meaningful for the research process for five main reasons:
1 They provide an overview of the material for larger research teams in which not
every team member can systematically work through every text (team aspect).
2 The summaries are a good starting point for creating tabular case overviews for
multiple cases (comparative aspect).
3 They highlight the differences between the individual cases (aspect of analytical
differentiation).
4 They help to generate hypotheses and categories (inspiring aspect).
5 They represent a first result of a case-based approach and can give the readers of
your research report a useful insight into the respective case, possibly in a form
enhanced by the later analysis (aspect of transparency and presentation).
Chapter objectives
In this chapter you will learn about:
The study was conducted by students within the ‘Environmental Education and
1
others, and their own positioning in a “global society”, cause a discrepancy between knowl-
edge and behaviour when it comes to protecting the climate?’ (Kuckartz, 2010b).
The sample consisted of 30 participants who were divided into two age groups: 15–25
years of age (‘network kids’) and 46–65 years of age (‘baby boomers’). The data consisted of
two parts: a qualitative, open survey in the form of an interview; and a standardized ques-
tionnaire that collected respondents’ social and demographic characteristics as well as their
general assessment of climate change using scales. We started with a problem-centred inter-
view (Witzel & Reiter, 2012), which followed the interview guide shown in Figure 5.1.
Figure 5.1 Excerpt from the interview guide of the example study
7) Write up results,
document process
Phase 1: Initial Work with the Text: Read, Write Memos and
Case Summaries
As with all forms of QCA, the first analysis steps consist of initial text work, writing
memos and preliminary case summaries. How this is done has already been described in
Section 4.5, so only a brief summary is given here. The beginning of the structuring QCA
process is marked by an engaged, careful reading of the text and the selection of particu-
larly important text passages to highlight. You can note comments and observations in
the margins and create memos to record things that strike you as particularly interesting
or relevant as well as any ideas you may have regarding the analysis. At the end of this
first phase of working through the text, you can write an initial, short case summary.
problems in the world, it is logical that ‘biggest problems in the world’ serves as a main
category for our analysis. The same is true for the topic ‘personal behaviour towards
climate change’: because this point is central to the entire research project, it is
included in the research question and serves as one of the main topics for the analysis.
It is possible that researchers may discover new, unexpected topics through their care-
ful reading of the text. It is best to approach working with the text as you would
approach open coding in a grounded theory approach: by writing notes regarding the
new topics in the margins and/or memos. As a rule, you should note anything that
seems relevant or peculiar at first; as you work through the data, you will gain a better
sense of how to distinguish between random topics and topics that could be signifi-
cant for the given analysis.
Whether you develop the categories and sub-categories – according to the proce-
dures described in Chapter 3 – directly using the data or deductively based on the
theoretical framework of the research questions or the interview guide, you should first
process some of the data in order to check if your categories and sub-categories and
their definitions can actually be applied to the empirical data. How much of the mate-
rial should be included in such a test depends on the size of the entire data set and the
complexity of the category system. The more complex the data and the greater the
number of categories, the more material should be included in the trial run. In general,
10–25% of the data should suffice for the initial test of the applicability of the topics
and categories. The trial run should facilitate a smooth transition to the next phase, and
the coding already performed initiates the first coding cycle of the entire material.
The demand for precisely defined categories in classical content analysis is often mis-
understood to imply that a text passage can only be assigned to one category. This is only
true for category systems that are designed in such a way that sub-categories exclude
each other (see the first example of deductive category development in Chapter 3). In
thematic coding, it is assumed that a given text passage can refer to multiple topics and
thus be assigned to multiple categories.
In our example project, we developed categories using the general thematic structure
of the interviews (Figure 5.3).
In addition to the general quality standards for good category systems from Section 2.4, the
following rules apply to the category system used in the first coding process. The category
system should:
• Be established in close connection to the research question and goals of the given project
• Not be too detailed or too broad
• Contain precise, detailed descriptions of the categories
• Be formulated with the report of results in mind – for example, categories should be
selected that are suited to give structure to the research report at the end of the analysis
process
• Be tested on a section of the data.
The entire data set is coded during the first coding cycle. Of course, this raises questions
about the scope of the segments to be coded: you must bear in mind their size. The fol-
lowing excerpt from an interview illustrates this issue.
I: In your opinion, what are the biggest problems in the world in the 21st century?
R1: Well, that’s a totally broad question … I would definitely say that religious and cultural
conflicts are some of the most difficult, and of course the environment and natural conflicts,
because, well, I believe you can’t really rate them because all the conflicts are affecting the
world and they are very deeply rooted … From conflicts about water to religious conflicts,
there are many, many conflicts! But I think the environment and cultural and religious con-
flicts are currently the most serious.
The passage clearly falls under the category of ‘WP – biggest problems in the world’. A
coded segment should be large enough that it can still be understood when it is taken
out of its original context. If the answers to a guideline-oriented interview are relatively
short, they can be coded quickly when the entire answer to a given question is coded
as a unit. The whole section pertaining to a given topic, which could contain multiple
paragraphs, should be assigned the relevant code, here ‘WP – biggest problems in the
world’. This approach prevents the same category from being assigned to the same sec-
tion or paragraph of text repeatedly. It is possible that other topics and categories could
be mentioned in the middle of this section, in which case those individual sentences
would also be assigned to a second category.
The following rules for coding – how to assign text passages to categories – can be
formulated:
1 The scope of a segment is not defined by formal criteria, but by semantic boundaries.
Thus, these units of meaning should always be complete thoughts and are mostly full
sentences.
2 If the meaning unit comprises several sentences or paragraphs, these are coded as
one coherent segment.
3 If interview questions or clarifications are essential for understanding the respondent’s
statement, they are included in the coding.
4 When assigning categories, it is important to develop a good sense of how much text
surrounding the relevant information should be coded. The most important criterion for
this is that the given passage can still be understood when taken out of context.
a technique in which two or more members of a research team code interviews inde-
pendently. This requires a category system where the categories and sub-categories
are well described and explained with examples. Consensual coding improves the
quality of the research project and the reliability of the coding.
In the first step, two or more coders code the data independently. Then, in the sec-
ond step, the coders sit together, sort through the codes and coded segments, and check
for similarities and differences. After discussing their reasoning, they should aim to find
a consensus regarding the most appropriate coding. By doing so, researchers can often
fine-tune their category definitions and codings, using the disputed text passages as
prime examples.
If they are unable to reach a consensus, they should call in more members of the
research team or discuss the disputed passages with the entire team. This process makes
differences in coding and analysis visible and can lead to constructive discussions within
the research team. Unlike any issues with coding agreement that may arise in the course
of a quantitative text analysis, where the main concern is to achieve a satisfactory coef-
ficient of intercoder reliability, the focus here is on clarifying any discrepancies as a
group and coming to a consensus. Consensual coding requires at least two different
researchers who are, ideally, involved in the coding process from the very beginning.
Thus, it is generally advisable to work with two or more coders who code the material
independently of each other. The category definitions will almost automatically become
more precise and the text assignments to categories more reliable, if the data are coded
by multiple researchers. However, it will not always be possible to work with multiple
coders – for example, if you are writing a master’s thesis or dissertation you often do not
have others to support you. If this is the case, researchers should carefully look to
improve explicit category definitions and prototypical examples where necessary. There
is little doubt that coding by only one person is sub-optimal and should be avoided.
There may be an exception if the coding scheme contains only a few well-described main
categories, as is the case with a transcript of a strictly structured interview. Here the coder
would not have to make any real decisions regarding the correct assignment of categories
since the answers always belong to the corresponding question in the interview guide
which is the basis for the coding scheme in the first coding cycle.
• Selecting the category that you would like to differentiate, that is, the category
for which you would like to create (new) sub-categories.
• Compiling all of the coded text segments that belong to this category into a list
or table. This is called text retrieval.
• Creating sub-categories based on the data according to the procedures described
in Section 3.2 on inductive category development. Add the new sub-categories to
a list, which is initially not ordered. When working in a team, each researcher
can be responsible for suggesting sub-categories for parts of the data. If you are in
a team of four and have conducted, for instance, 20 interviews, then each
researcher can work through five cases and make suggestions.
• Systematizing and ordering the list of sub-categories, identifying the relevant
dimensions, and, if necessary, grouping the sub-categories into more general and
more abstract sub-categories.
• Formulating definitions for the sub-categories and illustrating these using
prototypical quotes from the material.
Table 5.1 Definition of the sub-categories within the main category ‘biggest problems in
the world’
government should be role models’, ‘everyday routines get in the way’, ‘too
expensive’, ‘public infrastructure is insufficient’, and a residual category ‘other’.
4 Associated areas of behaviour. The fourth dimension was defined by the areas
of behaviour that the respondents named. This dimension partially overlaps with
the first dimension, ‘current behaviour’. The purpose for defining an independent
dimension was to establish which areas of behaviour were mentioned within the
context of climate behaviour, regardless of whether one is currently doing
something or would be willing to act (change one’s behaviour) in this area in the
future. The manner in which such sub-categories were formed here is similar to the
course of action outlined for the ‘biggest problems in the world’. The sub-categories
are consistent with the sub-categories of the current behaviour dimension; a precise
definition was formulated for each sub-category.
⇩
Biggest world problems Personal behaviour …
Person 1 Summary of Summary of …
Person 1’s text passages on Person 1’s text passages on
biggest world problems personal behaviour
Person 2 Summary of Summary of …
Person 2’s text passages on Person 2’s text passages on
biggest world problems personal behaviour
… … … …
Figure 5.4 Thematic cases-by-categories matrix as a starting point (top) for thematic case
summaries (bottom)
R2: No, I personally do not. I have to say that I’m pretty easy-going and
live quite comfortably and like I said, I am not really sure that we are
responsible for the drastic changes in the climate in the first place.
32 OK, first you’d have to make it really clear to me that we really are the
culprit. Maybe there has always been climate change in nature. I mean,
Germany used to have a thick layer of ice on the ground and, yes, the ice
has indeed melted, so the climate has again changed and that’s just how
things go, the climate is constantly changing. For me, it would have to be
really clear that we are the main reason that the climate is currently so
altered.
33–34 I: Yes, and what do you think now, for example, of conscientious
consumption? Meaning, for example, buying organic food or Fair Trade
clothing. Do you think that’s a good thing in principle? Would you buy
organic, perhaps, if you knew more about it? Or do you think organic is the
same as normal fruit and vegetables, so you don’t really see the point?
R2: Well, organic, I would buy some organic products really just for
my health because I don’t know how much that really influences the
environment. Um, Fair Trade clothing, I don’t know, I just buy the clothes I
like. So I wouldn’t necessarily buy something just because it is Fair Trade.
If it looks good, I might prefer to buy it, perhaps to give poorer people a
chance.
36 Yes, in principle, I know that. Things are, I don’t know, produced in the low-
wage countries where people make a measly salary and sold in Europe or
in general in the West, for a higher price. (…) Um, I don’t know if (…) that I
if I choose to stop [buying such things] it will change much.
38 Yes, I just think (…) that I am the type of person who (…), what would the
people down there do if they had no job at all. That’s what I think. And if
I don’t buy the clothes, then maybe they would not be able to work at all
and then they would still be worse off, even if they work for next to nothing
down there.
49–52 I: So if you were to see, for example, that the people around you are doing
their bit, for example, your friends or the people that you interact with the
most, would you change your opinion and participate and work to combat
climate change or would you still be virtually ‘the black sheep?’
I: Aha. So you feel rather strongly about it? I mean, you don’t think that
anyone could influence you to change your mind?
R2: Yeah, I don’t know. I think it would be a little like (…) I mean, sure,
there’s a known group dynamic, but (…) I don’t think I would do something
like that just because other people are if I weren’t completely convinced.
That proves that it isn’t really better.
Figure 5.5 Coded text passages as a starting point for thematic summaries
Let us look at an example in the thematic matrix in Figure 5.4: the ‘personal behav-
iour’ column for Person 2. This cell of the matrix is filled with the coded statements from
seven passages within the interview. The coded passages are presented in Figure 5.5, in
which the numbers of the first and last paragraphs pertaining to each statement are
listed in the first two columns. These statements were summarized as follows:
Person R2 makes no effort to consider the connection between his personal behaviour and
protecting the climate. The reasons he gives for this are that it is questionable whether
climate change is driven by human activities and it is also questionable whether one can
really change anything by, for example, buying Fair Trade products. Moreover, his own
convenience and comfort are the deciding factors. Even if many of the people around R2
are environmentally conscious, he would see no reason to change his behaviour. Potential
areas for behavioural change are identified in the subjunctive: save electricity, use public
transportation, ride a bike, and buy organic products (but only for health reasons).
It is not necessary to create case-related thematic summaries for every single topic and
sub-topic. You might certainly focus on the topics that you find especially relevant
and for which you would like to create comparative case overviews later in the analysis
process.
This approach to writing case-related thematic summaries has many advantages:
• It is systematic, not anecdotal, since all cases are handled in the same manner.
• The summaries are based on original statements; thus, they are literally grounded
in empirical data.
• The analysis covers all areas, because all of the data that pertain to a given topic
are included in the analysis.
• The analysis is flexible and dynamic; researchers can add, expand, or edit the
summaries at any point during the analysis process.
• The analysis is well documented and it is easy for other researchers to understand
which original statements led to which summaries.
• The thematically oriented summaries are a very good preparation for subsequent
forms of analysis, such as a detailed individual interpretation (‘within-case
analysis’) and a cross-case analysis (‘between-case analysis’) as well as for type-
building analysis.
• If the analysis is carried out using QDA software, connections are established
within the thematic structure between summaries and original data, which allows
quick access to both.
It is important to emphasize that the types of analysis listed in the figure and
explained below are options for analysis; it is neither necessary nor mandatory to apply
all of them in a project. Although the ‘category-based analysis of the main categories’
placed at 1 o’clock forms the prelude to the analysis in almost all studies, this is fol-
lowed by different forms of analysis depending on the project. While in one study the
focus is more on the cross-case, category-based analysis, in another the case-based
approach to the analysis may dominate and in a third both are treated equally.
Visualizations have a special function, because on the one hand they represent an inde-
pendent form of analysis, but on the other hand they are also used in many other types
of analysis, for example, in the form of tabular case overviews.
It can be helpful to first draw up an analysis plan that shows which research ques-
tions are to be answered with which types of analysis. Usually, it is desirable to arrange
a smooth transition between the current phase and the subsequent writing in phase 7,
because the results and findings provide the basis of the presentation. Therefore, it is
advisable to record all important results during this stage in memos or as preliminary
text blocks in the results report. Accordingly, in the following presentation of the types
of analysis, we also include considerations about writing up the results.
Impact on world
Case problems Learnability
R01 Problems have existed for Parents are the most
– Membership a very long time and are important authority. They
therefore not so easy to have to set an example.
– environmental NGO: no
solve. No belief in climate Media can also play a role,
– Gender: male change because humans but one should be critical.
– Occupation: no don’t have that much
influence. Nevertheless, we
have to conserve resources
because of environmental
protection.
(Continued)
Impact on world
Case problems Learnability
R05 Conflicts of faith have In early childhood, children
– Membership environmental existed since the Crusades, should already develop an
NGO: yes there’s not much you can do awareness of nature, e.g.,
about it. The same applies through environmental
– Gender: male
to capitalism, the greed projects in kindergarten. But
– Occupation: yes for power is part of human it’s not easy if the parents
nature. It’s different with don’t get involved. You have
climate change, which can to get them on board.
perhaps be influenced.
Miles and Huberman (1994, pp. 172–206), Schmidt (2010, pp. 481–482), Hopf and
Schmidt (1993, p. 16), and Kuckartz et al. (2008, pp. 52–53) present instructive explana-
tions of how to create case overviews. A case overview is very similar to a thematic
matrix; however, it only includes selected topics or categories. Usually, tabular case
overviews compare a selection of cases, or in the case of relatively small samples (n < 15)
all the cases, studied together with regard to characteristics that are considered particu-
larly relevant to the research questions.
Case overviews provide a good basis for an analysis. Thus, Hopf and Schmidt recommend:
The rows of a case overview can be arranged to suit the given analysis. For example, you
can arrange it so that individuals with similar characteristics appear next to each other
in the table. It may also be useful to indicate numbers and frequencies, for they reveal
even more information about the data, including whether a phenomenon or collection
of phenomena occur frequently or should be considered as unique cases. It must be
noted that the numbers here are fundamentally different from the numbers you would
use within the framework of statistical tests in studies with representative samples where
the goal is to identify standard, universal statements. However, scientific knowledge
about relationships does not require representative samples. If it did, a good deal of
medical and pharmaceutical research would be meaningless.
Overall, it can be stated that creating thematic summaries as described above in the
‘Write Case-Related Thematic Summaries (Optional Intermediate Step)’ section can be
an analytically very effective step, which enables the displays of results in a different
manner, because now not only original quotations, but also their condensed analytical
processing can be compared. Case summaries and comparisons are made possible by
this step in the analysis, especially because it would be impossible to create such over-
views using original quotations and passages from the text due to space limitations.
Moreover, the comprised and more abstract overviews have more analytical power and
evidence. The compilation of the summaries in tabular case overviews can also be used
as a starting point for further condensations and summaries. Maksutova (2021) gives an
instructive example of how the case-by-case summaries of a category can be aggregated
for groups of interviewees.
To create an in-depth interpretation of an individual case, you have to reread their tran-
script carefully, concentrating on a specific topic or question:
At the end, responses are compiled that refer to this individual case. Depending
on the question at hand, these responses can consist of detailed or concise
descriptions, contextual evidence of relationships, or theoretical conclusions.
In-depth interpretations can be used to verify existing hypotheses or
assumptions, to come to new theoretical conclusions, or to question, extend,
or modify the theoretical framework. The technique used depends on the
interpretation of the question and the respective tradition to which the
researchers feel connected, e.g., the hermeneutic or the psychoanalytic.
(Schmidt, 2010, pp. 482–483)
Usually, it is only a detailed analysis of one or a few selected cases and not a step of anal-
ysis to be carried out for all research participants. The rules for in-depth interpretation
of individual cases are not as strict as those for tabular summaries and interpretations
presented in the previous steps. Here, researchers are free to choose between different
models of interpretation such as hermeneutic or psychoanalytic-oriented techniques (as
noted above).
phase 1, but especially in the analysis phase 6, drafts and completed text sections on the
results can emerge, especially if you worked intensively with individual topics and
aspects for which sub-categories are developed and applied in phases 4 and 5. Generally,
you should consider phases 6 and 7 as an iterative process, because they may be run
through several times in succession for different topics: analysing and writing results for
topic 1, and again analysing and writing results for topic 2.
In category-based analysis, retrieving the texts that correspond to a given category or
sub-categories serves as the starting point for the writing process (see also Chapter 8 on
software support for this task). Take, for instance, the category ‘biggest problems in the
world’ in our example project. Let us examine the issues that the respondents indicated
as the biggest problems in the world today (which we defined as sub-categories). Which
problems are named frequently? Which are seldom named? Which problems are fre-
quently named in association with other problems? Which groups of respondents
named which problems? Examining a thematic category in detail often leads to the
development of additional sub-categories or dimensions. You do not always have to go
through the entire data set again in order to differentiate between these different
dimensions and recode the relevant text passages. You can also organize the answers in
a simple, systemized manner and prepare them for the research report.
When it comes to the research report, it is important that you define your framework
and determine how you would like to report the results of the analysis and how long
you would like your research report to be. For instance, it may be suitable to allot 60
pages to the results section of a dissertation, but you would only be allotted 5–10 pages
for an article in a journal or book.
Developing a storyline is quite useful. Starting with the research questions that serve
as the basis for the research, you can develop the story by introducing the categories
that you wish to write about in such a way that they pique the reader’s interest. If the
study was theory- or hypothesis-oriented, the corresponding categories and concepts
should of course be considered first when developing the structure. Once you have
conceptualized the general structure of the research report, you can adjust the number
of pages allocated as necessary, but it is helpful to know what you are aiming for as you
begin writing about your analysis of the first category.
At the end of each report on the results, the original research questions should be
addressed once again. Could these questions be fully answered by the study? Were any
assumptions and hypotheses confirmed or refuted? Which questions could not be
answered with the data collected? Where can (knowledge) gaps be identified? What new
questions arose during the research process?
large to include directly in the report, it should be included in the appendix at the end
of the report. Coding rules and prototypical examples should also be documented in the
appendix, perhaps via a QR code containing a link to a website. Examples of coding
rules and prototypical examples for some of the categories can be included in the meth-
ods section to illustrate the methodological approach. Further notes on documenting
the procedure and writing the results report can be found in Section 9.6.
Chapter objectives
In this chapter you will learn about:
In structuring QCA, categories are created using topics and sub-topics, which impacts
the following phases of the analysis. Evaluative qualitative analysis is different, as can
be seen in Figure 6.1, which presents the typical evaluative analysis process for a sin-
gle evaluative category. When analysing multiple categories, simply proceed through
phases 2–5 for each category.
7) Write up results,
document process
Phase 1: Initial Work with the Text: Read, Write Memos and
Case Summaries
As with all three types of QCA, the evaluative analysis begins with the task of familiar-
izing oneself with the material – initial work with the text, writing memos, and writing
case summaries, as described in detail in Section 4.5. The material is read carefully and
commented on, and important findings are recorded in memos. If it is already clear at
the beginning of the analysis that evaluative categories are to be developed, then atten-
tion can be paid to the relevant aspects: Which topics or themes would be suitable for
an evaluative analysis? Which cases can be characterized as special, possibly extreme
cases regarding selected features? And when writing case summaries, any evaluative
assessments already made can be included in the summaries.
to code, the same considerations apply as in structuring QCA (see p. 103−105). If a category
has already been coded thematically, you can build on previous codings and save time by
skipping this phase. One evaluative category could conceivably integrate multiple existing
thematic categories.
not to select respondents or groups of respondents systematically, that is, do not select
only men or interviewees belonging to a specific socio-demographic group. The more
heterogeneous and the more difficult the data set is to evaluate, the more cases must
be included.
If your data set is composed of distinct groups (such as the age groups 15–25 and
45–55), you should include a representative selection from each group, such as five from
the younger group and five from the older group. In most other cases, a random selec-
tion would be better.
In essence, the first way of defining evaluative categories consists of the dichotomous
characteristics ‘sense of responsibility exists’ and ‘no sense of responsibility’. The third
characteristic is allotted for ambiguous or unclassifiable cases or respondents. The
advantage of using this minimalistic variation with only three characteristics is that
there is only one differentiation that has to be made, which can be articulated by precise
definitions and fitting prototypical examples. The disadvantage is that the less differen-
tiated assessment of the data will lead to limitations in subsequent phases of the
analysis. In contrast, selecting five different characteristics produces a more detailed
assessment (Table 6.2).
*This table and Table 6.3 only contain a few selected citations as examples. For space reasons,
we have omitted further source information regarding the prototypical examples here.
While testing the applicability of these characteristics on our data, it became evident
that ‘high sense of responsibility’ was never assigned because the coding rules were so
strict. The same is true for the fourth characteristic, ‘no sense of responsibility’, which
was never assigned since it could only be coded if a person specifically indicated feeling
‘no sense of responsibility’.
Given the relatively small number of interviews to analyse (n = 30), we pragmatically
decided to distinguish only between three characteristics plus the characteristic ‘unable
to classify sense of responsibility’ if there is insufficient information. The characteristics
were then defined as shown in Table 6.3 and provided with prototypical examples.
using additional examples. Any uncertain cases should be discussed with other research-
ers (e.g., in your working group or in a colloquium for junior researchers).
Reminder: if there is more than one evaluative category, go through phases 2–5 for
each of them.
B. Qualitative-interpretative
F. Quantitative relationships between analysis of evaluative categories
Analysis Options
two evaluative categories (crosstab)
C. Tabular case overviews
• Presentation of what was said in what way and using which arguments (e.g.,
regarding the topic of responsibility), sorted by characteristic.
• Presentation of general and extraordinary statements, the latter of which are
usually found in marginal characteristics. For example, the person who is
completely ignorant regarding climate change is in the group ‘People with little
or no sense of responsibility’.
Biggest WP:
nature and Sense of Statements regarding
Case environment Age group responsibility personal behaviour
Person 1 no 15–25 low ‘Yeah, basically, I know
that. … I don’t think much
will change if I stop.’
Person 2 yes 46–65 low ‘I think I should be fine
for the rest of my time on
this planet …. Otherwise, I
think it doesn’t really matter
to nature whether there are
people or not.’
Person 3 yes 15–25 high Intentionally
environmentally conscious.
However, ‘sometimes my
hands are tied for financial
reasons’.
Person 4 no 15–25 low Throws little trash away
(i.e., recycles); does not
drive a ‘gas-guzzler’.
Table 6.5 Quote matrix for the relationship between an evaluative and a thematic
category
Sense of responsibility
High Moderate Low
Food/diet Text passages Text passages Text passages
pertaining to the pertaining to the pertaining to the
eating habits of eating habits of eating habits
people with a people with a of people with
high sense of moderate sense of a low sense of
responsibility responsibility responsibility
Table 6.6 Crosstab for the relationship between two evaluative categories
Sense of responsibility
Personal
behaviour
to protect
environment High Moderate Low Total
Yes Number of Number of Number of Total number
people with people with people with of people
positive personal positive personal positive personal with positive
behaviours and behaviours and a behaviours and personal
a high sense of moderate sense a low sense of behaviours
responsibility of responsibility responsibility
No Number of Number of Number of Total number
people with people with people with of people
negative personal negative personal negative personal with negative
behaviours and behaviours and a behaviours and personal
a high sense of moderate sense a low sense of behaviours
responsibility of responsibility responsibility
Total Total number of Total number of Total number
people with a people with a of people with
high sense of moderate sense a low sense of
responsibility of responsibility responsibility
Table 6.7 Crosstab for the relationship between an evaluative category and a
socio-demographic variable
Gender
Sense of
responsibility Male Female Total
High Number of males Number of females Total number of
with a high sense of with a high sense of people with a high
responsibility responsibility sense of responsibility
Gender
Sense of
responsibility Male Female Total
Moderate Number of males with Number of females with Total number of
a moderate sense of a moderate sense of people with a
responsibility responsibility moderate sense of
responsibility
Low Number of males Number of females Total number of
with a low sense of with a low sense of people with a low
responsibility responsibility sense of responsibility
Total Total number of males Total number of females
gives the cases an overall score. It is questionable if this is more effective in practice, for
researchers often like to (or must) include broader context in order to evaluate text pas-
sages correctly. From a hermeneutical standpoint, it would be difficult to justify not
taking the entire text into consideration and simply focusing on the statements directly
surrounding the given text passage. However, by including the broader context, you can
assess the entire text in relation to the given category.
In general, the categories in the evaluative analysis are more holistic than the catego-
ries or sub-categories in a content-structuring QCA. The classifications and assessments
to be made in evaluative QCA demand more of the coders than in structuring analysis.
It is hard to imagine that multiple coders could reach agreement without having spe-
cialized knowledge in the field of research. Coders must understand what they are doing
and be able to justify their codings using the data. We recommend having at least two
coders working independently in evaluative analysis. Of course, there may be situations
in which analysts must work alone, such as when writing a thesis or dissertation; how-
ever, even then it can be beneficial to think about when a second person should be
brought in to assist you or double-check your work.
The evaluative approach is especially suitable for theory-oriented research. This does
not necessarily mean that you must have profound theoretical knowledge regarding the
research question and seek to investigate explicit hypotheses, as it is quite conceivable
that an interest in the formulation of hypotheses and theories only develops during the
course of the project. However, evaluative qualitative analysis is especially suitable for
theory-oriented research; if the primary aim is description, the method of structuring
content analysis is usually more appropriate.
Of course, it is also possible to combine evaluative QCA with structuring analysis
and define evaluative categories only for particularly interesting themes or topics. In
some cases, the evaluative categories could build on the content-structuring coding
and make use of the work already done on the text. Evaluative QCA is a method that
clearly demonstrates that the frequently expressed characterizations of qualitative
research as mainly explorative, for generating theories, and not suitable for testing
theories, are too restrictive. Those characterizations describe the mainstream of qualita-
tive research, but they are not always true without exception. Hopf et al. (1995) argue
that standardized questionnaires with predetermined answers are often inadequate in
social research because they cannot express the complexity of the research question.
So, in theory-testing research too, it may be useful to let respondents answer in their
own words with whatever nuances they may express and then analyse their responses
as a research team and try to reconstruct the levels of the evaluative category or catego-
ries. This approach often yields more valid information and data than standardized
tests and instruments.
Also in evaluative qualitative analysis, the analysts can follow the case-oriented perspec-
tive in the form of in-depth interpretations of individual cases. In addition, crosstabs may
present an overview of the data using thematic as well as evaluative categories; however,
they cannot be compared with those based on a large number of representative studies.
Chapter objectives
In this chapter you will learn about:
The type-building QCA presented in this chapter shows a way to progress from structur-
ing QCA and/or evaluative QCA to the construction of typologies.
According to many qualitative methodologists, creating types and developing a
typology are the main goal of qualitative data analysis (Creswell & Plano Clark, 2011,
pp. 212–238; Kluge, 1999; Lamnek & Krell, 2016, pp. 218–227; Schmidt, 2000). Type
building is an excellent method for analysing target groups and makes it possible to
identify and analyse multidimensional patterns in the data. Metaphorically speaking, it
is a matter of systematizing and analysing the cases-by-categories matrix (as shown, for
example, in Figure 4.2), as a systematization and meaningful ordering of the cases in
their entirety. With the help of qualitative content analysis, the developing of types is
possible in a methodically controlled form. Compared to structuring and evaluative
QCA, type-building analysis is more complex and methodologically more demanding.
For this reason, the characterization of the procedure in this chapter starts with a discus-
sion of its methodological foundations.
The real core of type building involves searching for multidimensional patterns and
models that enable researchers to understand a complex subject or field. Type-building
analysis often builds on the preliminary work done in structuring or evaluative QCA.
1
Approaches to building types can often be found in biography research, in youth
research, in lifestyle research, and in interdisciplinary research fields such as public
health and environmental awareness and attitudes.
2
This study was conducted towards the beginning of the 1930s by the Research Unit for
Economic Psychology at the University of Vienna under the direction of Paul Lazarsfeld and
Marie Jahoda. See http://agso.uni-graz.at/marienthal/e/pictures/15_marienthal_study.htm
(accessed 20 March 2022).
By constantly comparing and contrasting the individual cases according to the above cri-
teria, researchers were able to identify four different mindsets for dealing with unemploy-
ment and responding to the deprivation it caused (Jahoda et al., 1933/2002, pp. 64–82):
• The unbroken continue their daily life and household, search for a new job, and
stay active and happy.
• The resigned maintain their daily life and household but do as little as possible
and refrain from making future plans.
• People in despair have lost hope and are moving backwards; they make little
effort to improve their situation and do not even bother to look for work.
• The apathetic have lost the energy to take care of their home and their children;
they are like passive bystanders to what is happening to them and do not even
try to change their situation.
These mindsets were precisely based on the data and their characteristic features were named
explicitly, as can be seen in the following excerpt from the description of the ‘resigned’:
If we were to single out from this description the criteria which lead us to
categorize a family as resigned and summarize them epigrammatically, we
would say: no plans, no relation to the future, no hopes, extreme restriction of
all needs beyond the bare necessities, yet at the same time maintenance of the
household, care of the children, and an overall feeling of relative well-being.
(Jahoda et al., 1933/2002, p. 53)
Looking back at the history and development of empirical social research, we find not
only practical applications for building types, but also works that reflect on the method-
ical foundations of type building, such as Weber (1964), Hempel and Oppenheim
(1936), Lazarsfeld (1972), Schutz (1962), Kluge (1999), Kelle und Kluge (2010), Bailey
(1973, 1994), and Kuckartz (1991). Schutz examined everyday life and concluded that
‘the individual’s common-sense knowledge of the world is a system of constructs of its
typicality’ (Schutz, 1962, p. 7). According to Schutz, all of the knowledge that comes
with experience is organized in the form of typical experiences. We do not experience
our surroundings and environment ‘as an arrangement of individual unique objects,
dispersed in space and time, but as “mountains”, “trees”, “animals”, “fellow-men”’
(Schutz, 1962, pp. 8–9). Types are often constructed in anthropology, though the basic
aim there is to understand types in a psychological sense of understanding individuals’
inner lives. In social science research, the goal of analysis is simply to understand what
is typical. Schutz follows the tradition of Max Weber, who declared that constructing
comprehensive types was the main goal of empirical social science research. Types as
analytic devices link hermeneutic methodology, the aim of which is to understand
individual cases, with social science statistics, the aim of which is to find standard inter-
relationships and correlations (Kuckartz, 2006; Lazarsfeld, 1972).
10 social milieus can be selected for a given household. Regardless of the method used, each
type is – implicitly or explicitly – based on the notion of an attribute space.
In both qualitative and quantitative research, the process for building empirical types
contains five phases (Figure 7.1):
1 Determine the attribute space. In this phase of the analysis, researchers must define
the attribute space (i.e., select the categories/attributes by which all cases are to be
compared) that serves as the basis for type building.
2 Group individual cases and build typology. In this phase, typologies are constructed,
that is, cases are grouped into clusters (types). After comparing and contrasting
the typologies, you can decide which typology is best suited to the data. This
phase is thus an experimental phase. Several solutions for grouping the individual
cases are tested and compared.
3 Describe typology. The constructed typology and the individual types that have
been created are described in greater detail.
4 Explicitly assign individual cases to the types created. This phase shifts the focus from
the groupings back to the individual elements. Here, individual cases (usually
respondents) are assigned finally to the created types.
5 Analyse relationships. In the final phase, the typology and the different types
within it are presented according to their characteristics and the relationships
between the types and secondary variables are analysed.
5) Analyse
relationships
4) Assign cases
to types
1) Determine the
attribute space
Research
questions
3) Describe
typology 2) Group individual cases
and build typology
The first phase of building types involves deciding which attributes are relevant for
the desired typology and determining what information is available in the data that
have been collected. In the above example of mindsets from the Marienthal study, all
of the attributes that were used to describe the families were relevant to the typologies.
The number of attributes you can include in your type building depends on the con-
struction of the typology. There are three main approaches to constructing typologies:
Environmental behaviour
positive negative
Environmental high Type 1: Consequent Type 2: Environmental
awareness protector rhetorician
low Type 3: Uncommitted Type 4: Ignorant to
protector environment
Two of the types created, specifically types 1 and 5, are homogeneous in terms of attri-
butes, meaning that both parents of all of the respondents within the type have the same
level of education. In the case of type 1, both parents graduated from college; in the case
of type 5, neither parent finished school. The other types created via reduction (types 2,
3, and 4) show variance, for they contain respondents with different attributes, in this
case different levels of parental education. For example, five different attributes can be
assigned to type 3. The determining factor is that one parent graduated from high school;
the other parent could have graduated from high school, middle school, or never have
finished school.
the other hand, are built directly using the empirical data, meaning that respondents are
grouped according to types which are as homogeneous as possible internally and as het-
erogeneous as possible externally. Such types are almost always polythetic; the individuals
that belong to a type are not absolutely the same in terms of the attributes within the
attribute space, but they are quite similar.
Natural typologies can be ordered according to systematic, intellectual structures as
well as using statistical algorithms. Cluster analysis methods are especially suitable for
the latter (Kuckartz, 2010a, pp. 227–246). A good way to build complex polythetic types
without such formal algorithms is to systematically process and group case summaries,
as presented in the five phases shown in Figure 7.2.
Phase Task
1 Define attribute space and create a case summary for each respondent that is focused
on these attributes
2 Sort, order, and group case summaries according to similarity
3 Decide on a reasonable number of types to be built
4 Formulate creative names for each type that poignantly express the main
characteristics of each type
5 Assign each respondent to a type; order the respondents according to their similarity
to the central type
In research practice, you can implement this method by writing each case summary
on a note card and asking the research team to arrange these cards (which represent
respondents) on a large bulletin board. To do so, proceed as follows:
1 Distribute the case summaries evenly among members of the research team.
2 Each researcher must now closely review the cases that have been assigned to him
or her, checking the case summary and revising it as necessary before writing it
on the note card.
Working as a group
Constructing, arranging, and assigning an order to types in this way represents a cre-
ative act and will not produce a precise, canonized description. Here are a few helpful
tips. It is often useful to work through this type of categorizing and typifying task as a
group. The technique described of writing case summaries on note cards and grouping
them together has proven successful in research practice, but this does not mean that
successful type building is only possible in a team. Even when working alone, it can be
useful to proceed in a similarly methodologically controlled way.
9) Write up results,
document process
Marienthal study, the mindset typology is especially strong because the distinction
between the four types is clear, plausible, and comprehensible. If the researchers had
attempted to differentiate between eight or more types, the unions and politicians – the
main recipients of the study – would likely have had more trouble understanding the
results. Conversely, merely including four social milieus in a lifestyle research project
would likely be perceived as too simple because current lifestyle models, such as the
Sinus Milieus, include 10 or more different types.4
In order to determine how to build your types, you must consider your sample size
and the dimensionality of your desired attribute space:
• Monothetic homogeneous attribute types can be created using two or three attributes
with relatively few characteristics.
• Building types by reduction is much more flexible and can include more features
and more characteristics.
• Polythetic type-building, however, makes it possible to integrate numerous
characteristics and define a truly multidimensional attribute space.
4
The last version of the Sinus Milieus was developed in 2022 and differentiates between
10 significant clusters (milieus). See www.sinus-institut.de/en/ (accessed 20 March 2022).
analysis techniques, such as text retrieval, are very effective aids in this part of the analysis
process. In the Marienthal study, this strategy of representative case interpretation would mean
selecting a person who is the best example of the ‘unbroken’ type. This person and his or
her characteristics would then be presented and interpreted. The same would be done for
the other three types, the ‘resigned’, those ‘in despair’, and the ‘apathetic’. Select the most
representative persons and describe them in detail. The result is that four – real – people are
shown as examples to produce a better understanding of the constructed typology.
The second option for selecting which cases to study with in-depth analysis involves
constructing a model case based on a synopsis or montage of the most suitable text segments.
This approach is less focused on the individual cases and in many ways is similar to Weber’s
approach of building ideal types (Kuckartz, 1991). However, because the polythetic types
and their position in the attribute space have already been determined via the typological
interpretation of the text, they are real, not ideal types, and they represent real individual
respondents from the sample. By reviewing relevant text passages, you can determine
which cases would be suitable for the given type and include them in the synopsis.
coded in addition to the typology (e.g., with thematic or evaluative categories). If this
did not happen, then this phase is omitted.
If parts of the material have been structured with the use of further categories (the-
matic or other categories), then connections to these categories can now be analysed. For
the cases assigned to a type, the segments coded to specific thematic categories can be
compiled and compared with those for other types. Similarly to the evaluative categories
described in Section 6.2, a quote matrix can be created for comparison. Statistical analyses
of the correlations are also possible: using cross-tabulations, it can be shown whether and
how often certain topics were coded for each type or for how many members of a type.
Formally, the identification of a type is similar to an evaluative category as described
in Chapter 6. For each individual, the type to which he or she belongs is stored. In this
respect, all kinds of complex analyses described above for evaluative content analysis
are also possible for a typology, for example the analysis of relationships to other evalua-
tive categories. Since evaluative categories are a classification on the one hand, and are
based on specific text passages on the other, relationships can be analysed both qualita-
tively and quantitatively. This means that both simple and complex statistical analyses
can be calculated, and the corresponding text segments or thematic summaries can be
compiled for the groups formed.
strong
The Rooted Woman . . . The Collector . . .
weak
Chapter objectives
In this chapter you will learn about:
Under the name ‘QDA software’, a special type of computer program is available that is now
popularly used in qualitative research. For over three decades, the field of computer-assisted
analysis of qualitative data has been considered one of the most innovative fields of social
science methodology development. QDA software programs do not prescribe one specific
method of analysis but can be used for many types of data and in a variety of methodolog-
ical approaches (Kuckartz & Rädiker, 2019; Silver & Lewins, 2014). For example, Creswell
and Poth (2018, pp. 207–218) describe how QDA software can be used in five different
research traditions, including biographical life history, phenomenology, grounded theory,
ethnography, and case study.
In this chapter, the capabilities of QDA software are presented, focusing on qualita-
tive content analysis. The aim of this chapter is to give an overview of the different
possibilities for analysis. Since software quickly becomes outdated and new versions are
usually accompanied by new interfaces and changed menu navigation, this presenta-
tion focuses on the general procedures and does without specific instructions in the
sense of ‘select menu item XY’, ‘click here’, and ‘double-click there’. Detailed overviews
of QDA software’s options and features can be found in Silver and Lewins (2014), Friese
(2019), Jackson and Bazeley (2019), and Kuckartz and Rädiker (2019).
At the beginning of any kind of QCA, questions arise regarding how best to manage
and organize the data. How should we format the data? How can the data be analysed
using QDA software? How should we organize, save, and store files and folders? How
can we organize and coordinate the work within a research team?
5 Format the transcriptions in such a way that the QDA program can be used
optimally.
6 Save and archive the transcriptions as RTF or DOC/X files.
7 Import these files into the QDA software.
The first three steps are of course only necessary if new data material has been collected
and must be transcribed. If the text data has already been digitalized, proceed with step 4
to anonymize, format, and import the existing data into the QDA software.
If possible, when working with qualitative interviews, group discussions, focus groups,
and similar forms of data collection, you should work with audio recordings rather than
with manuscript notes to jog your memory.
Table 8.1 presents the advantages and disadvantages of using audio recordings. The
advantages of audio recording are obvious, unless you are dealing with particularly
sensitive issues which require a very confidential interview setting that would be dis-
turbed by simultaneous audio recording. Writing an exact transcription is only possible
using an audio or video recording, meaning that such recordings are the only way to
ensure that you can use direct quotations in later steps of the analysis as well as in the
research report.
Audio recordings are best carried out with digital recording devices; usually a smart-
phone offers quite sufficient quality for face-to-face interviews.
The same is true for video recordings, though researchers usually focus on the
audio recording of the interview. Modern smartphones, digital recorders or video
cameras, and transcription software make recording and transcribing much easier
than in the past. The rules for transcription presented below apply to both audio and
video recordings.
1
Further references to more complex transcription rules can be found in Kuckartz (2010a,
pp. 38–47).
4. Language and punctuation should be smoothed slightly to +/– (no approximation to written
accommodate written standards. language)
5. Long, clear pauses should be marked by three full stops in – (recognizable by time stamps,
brackets ‘(…)’. if applicable)
Figure 8.1 Transcription rules for computer-assisted analysis and notes on their
implementation in automatic transcription software
Non-Automatic Transcription
For the ‘manual’ transcription of audio and video files on the computer, a number of
specialized programs such as the following software tools are available, some of which
are even free of charge:
• EasyTranscript (www.e-werkzeug.eu/index.php/en/)
• ExpressScribe (www.nch.com.au/scribe/)
• f4/f5transcript (www.audiotranskription.de/en/)
• HyperTRANSCRIBE (www.researchware.com)
• InqScribe (www.inqscribe.com)
• Transcriber (https://sourceforge.net/projects/trans)
It is also possible to transcribe directly with QDA software such as MAXQDA, ATLAS.ti,
or NVivo. Most transcription programs are easy to use and contain all of the functions
normally required for transcribing interviews within social research. As with most media
players, you can play, stop, pause, restart, rewind, and fast-forward as desired. In addi-
tion, you can adjust the playback speed as well as the return interval, that is, the time in
seconds that the recording rewinds and resets upon stopping. A footswitch or pedal is
useful for stopping and starting the recording while transcribing.
The transcription software can insert a time stamp into the transcript, for instance at
the beginning or end of a paragraph (when you press the Enter key). This makes it pos-
sible to synchronize the text and audio recording within the QDA software so that you
can click on a given time stamp when reading the text in order to play that exact
excerpt from the audio recording. The transcription rules described above can thus be
implemented without any problems.
Figure 8.2 shows an excerpt from a transcript that was created in accordance with the
rules presented above (excluding time stamps).
R7: My boyfriend and I have a study group. I mean, I explain everything to him two times and then I
understand it better myself. (…) And, yeah, I studied with one of the other students from my statistics
group one time, too.
zI: And how do you feel when you are studying? Do you have a positive or negative attitude or feeling
towards statistics?
R7: I like it a lot. I didn’t think that would be the case, but I always liked maths, so I think that’s why I think
it’s OK.
I: And did that change at all over the course of the semester?
R7: Yes.
I: And how?
Automatic Transcription
Recently, there has been great progress in the field of automatic transcription. What
seemed a long way off a few years ago is now state-of-the-art: if audio and video files are
available in good recording quality and the recorded voices are as dialect-free as possible,
they can be automatically converted into a transcript with the help of artificial intelligence.
In the process, time stamps are also inserted into the text, the speakers are identified, and
their conversational contributions are marked accordingly in the transcript. In this way,
even transcripts of focus groups with different speakers can be created automatically.
Even though the market of providers and the functionality offered are subject to fast
change, we would like to mention here some providers that have become better known,
and we have also added the location of the company as a first clue regarding data pro-
tection: AmberScript (Netherlands), f4x (Germany), HappyScribe (Ireland), Otter.ai
(USA), TEMI (USA), Trint (UK), Sonix (USA). Some of these software programs allow you
to edit the reference dictionary to increase the recognition rate. The biggest difference
between the providers, however, concerns the languages that are supported. While
some companies offer only English, others allow you to choose from more than ten
different languages. In the meantime, Microsoft Teams, Zoom.us, Cisco WebEx, and
other tools for conducting video conferences also offer the possibility of transcribing
conversations live and inserting live subtitles. After an interview carried out with these
tools, in addition to the audio and video recording, a transcript can be downloaded that
then can be imported into QDA software.
Although the quality of automatically generated transcripts is steadily improving,
one should not expect the technology to work error-free. Automatically created tran-
scripts always need to be checked and corrected manually. Instead of 5–10 hours for
manual transcription, only about 3–6 hours are necessary to check and correct the
transcript of a one-hour interview. In addition to the time saved, there is also an ana-
lytical gain: if the transcript is checked directly in the QDA software, the initiating work
with the text can also be started during the process of checking the transcription.
With all the possibilities offered by automatic transcription, however, two points
should not be ignored. Firstly, it is important to pay attention to the aspect of data
protection and privacy because, although typically confidential, the audio and video
recordings usually have to be uploaded to the providers’ servers. Therefore, the server
locations especially, as well as the possibility of concluding contracts for commissioned
data processing, have to be checked before choosing a transcription service provider.
Secondly, it should be taken into account that most providers do not specialize in social
science transcripts and the automatic techniques may only capture some of the ana-
lytically relevant information. Even the simplest information, such as pauses in the flow
of speech, is rarely represented, which is why, in our proposed transcription rules in
Figure 8.1, we have pointed out corresponding limitations in automatic transcription.
It must be decided, having regard to the research questions, whether the limited accu-
racy of automatic transcription is acceptable for a proposed research project.
Regardless of whether manual or automatic transcription is used, and regardless of
which transcription system and software is used, the transcripts should be formatted in
such a way that they are easy to read on the computer screen in later steps of the analy-
sis process and that they can utilize various functions within the QDA software programs,
particularly the lexical or word-based search options. For example, it is important that
the terms used to identify either the speaker, a specific question from the interview guide,
or a particular section of the survey are used consistently throughout the entire text. You
should choose how you would like to abbreviate each speaker’s name and use the same
abbreviation throughout the entire work, such as ‘I’ or ‘INT’, and not interchangeably ‘I’,
‘INT’, and even ‘Interviewer’. Uniform spelling and referencing are absolutely imperative
if you plan to use the lexical search options in QDA software later.
After the transcription is finished, the entire text should be proof-read again; errors
should be corrected if necessary. It is recommended that interviewers take the time to
compare the transcript with the audio recording one last time.
In qualitative research, the analysis of the interview begins with or even before the
transcription. Researchers develop ideas and perhaps even hypotheses during the inter-
view and while listening to the audio recordings. They keep the interview situation and
any particularities in mind, and may discuss them with the research team. All of these
thoughts deserve to be recorded, though not within the transcript itself. Such ideas
should be documented in the form of notes and memos that can be saved together with
the text and linked to the corresponding passages. In many cases, it has also proved
useful to prepare interview documentation, sometimes called a postscript, as soon as
possible after the interview (Witzel & Reiter, 2012). This postscript should include eve-
rything relevant: location, time, contact information and everything the researcher
noticed during the interview, such as communication behaviour, the course of the
interview, information on the interview situation, and external influences, should be
documented (Helfferich, 2011, pp. 193, 201).
originates, for example, Mehmet can become Kamil or Nadine can become Juliette’
(Przyborski & Wohlrab-Sahr, 2014, p. 170). Frivolous names such as ‘Donald Duck’
should also be avoided.
Place names, calendar dates, and the like should also be changed in such a way that
direct conclusions about research participants are no longer possible. Places can be
replaced by more general references such as ‘small town’ or ‘village’ and dates can be
changed to indicate general time frames, such as ‘summer’ or ‘last winter’. A table should
be created to summarize the changes and provide information for decrypting the
anonymization, which, of course, must be stored separately from the data in order to
maintain privacy and confidentiality. The storage of this information as well as the data
set itself must comply with the appropriate data protection regulations. The open-source
software QualiAnon, provided by the Qualiservice of the University of Bremen, Germany
(www.qualiservice.org), has been specially developed for anonymizing interview data and
allows the sometimes time-consuming process to be systematized and simplified.
After the data have been edited, formatted, and made anonymous, they should be
saved as an RTF or DOCX file and additionally backed up in a second location (e.g., in
the cloud). This is the only way to effectively protect against data loss, because a copy
on the same hard disk is of little help in the event of a disk crash.
Ideas that you noted during your first reading of the material or during the transcrip-
tion process must also be organized. You must decide if you would like to label them as
a special kind of text that you can link to the interviews or organize them as memos
that belong to a given text or text passage. The latter is advantageous because such
memos are always directly associated with the text that they reference and so are easily
accessible. The same considerations apply to the documents made by the interviewers
about the interview process (‘postscript’). These should be set up as memos attached to
the related transcript documents and not organized as separate documents; this way
they always remain associated with the interview and are easily accessible. Some
researchers prefer to have the documentation accompanying the interview (such as
postscript, interview transcript, etc.) available within the transcription as well. This is
also possible, but the argument against this is that then an analysis of the word frequen-
cies of an interview is distorted or additional selections are necessary.
The question of whether to subdivide the text into text units or units of meaning
goes back to the transcription phase but can only be answered later in the analysis pro-
cess. If you plan to code the data according to syntactic or semantic aspects in
subsequent codings, now is the time to split up the data into logical paragraphs. It is
useful to begin a new paragraph for each meaningful unit of text. In primarily interpre-
tive analyses, defining different groups in advance may seem unnecessary; however, the
more elements of QCA you plan to integrate into your analysis, the more useful these
sorts of subdivisions can be.
When different members of a research team are involved in analysing the data, it is
important to think about how to organize and coordinate your work as a team. Consider
the following questions:
programs in the original layout. When analysing studies based on interviews, DOCX or
RTF formats are more suitable than PDF, because they allow you to add and edit text at
any point in the analysis process. This might, for example, be necessary in order to
anonymize the text later. Moreover, only DOCX or RTF formats allow you to use a syn-
chronous time stamp in the transcription to access the audio and video files.
Some types of text (e.g., as transcriptions of focus groups, answers to open questions
in online surveys, and texts that stem from internet forums) can be imported into QDA
software in a pre-structured format. The software then automatically assigns text pas-
sages to categories and sub-categories or, in the case of group discussions or focus
groups, to different speakers during the import (Kuckartz, 2010a, pp. 49–55; Kuckartz &
Rädiker, 2019).
The term ‘memo’ for such notes, comments, and recordings is used in particular in
the context of grounded theory, in which memos play a central role. According to the
grounded theory approach, you should differentiate between different types of memos,
including memos that are theoretically oriented (‘theory memos’), memos that contain
the definitions of the categories and their characteristics or attributes (‘code memos’),
memos that refer to specific features of the language (‘linguistic memos’), and memos
that contain case summaries (‘document memos’). The latter can also be used for organ-
izational purposes, for example to record the analysis status of an interview text.
In QDA software, memos can be linked to any kind of object, such as whole texts,
text passages, categories, and sub-categories. It is recommended that you use different
symbols to represent different types of memos. Over the course of the analysis process,
you can combine individual memos to form larger, integrated memos that serve as
building blocks for the research report.
Category Sub-categories
Environmental behaviour • Travelling
• Recycling
• Energy
• Consumption
• Other
In principle, a category system can have any number of tiers: not only sub-categories,
but also sub-sub-categories and more levels are possible. Most QDA programs support
such a hierarchical organization of categories. However, it is usually recommended to
work with no more than three or four different levels. In the case of the category ‘envi-
ronmental behaviour’ above, a third level would be necessary if valences were to be
taken into account when coding. This would result, for example, in the categories ‘envi-
ronmental behaviour → travelling → positive’ and ‘environmental behaviour →
travelling → negative’.
In quantitative content analysis, the categories are often numbered consecutively, for
example as follows:
10 Environmental behaviour
11 Travelling
12 Recycling
13 Energy
14 Consumption
15 Other
Such numbering is not actually necessary in QCA, yet it is often practised; however, it is
problematic when new categories are developed and inserted into the category system
in the course of the coding process.
Figure 8.3 Interview excerpt with open codes displayed on the left
I: And would you like to do more than what you are already doing?
R29: Theoretically, yes, but the question is how. And I don’t know if I would feel comfortable in some
(…) We Save the World Association.
I: Do you feel a responsibility to address the problems of the 21st century?
R29: Personally yes, but not globally.
I: Can you clarify what you mean by that?
The development of categories is a lengthy process which requires that you go through
the data or parts of the data repeatedly. An alternative way of developing categories directly
from the text is the technique of category development via focused summaries, which is
described in Section 3.2. This procedure is close to the text in that the text passages that are
important for their content are first abstracted and paraphrased in a process that may involve
several stages. The procedure is quite time-consuming, but proves quite helpful, especially for
beginners. In a first step, QDA software allows paraphrases to be recorded directly next to the
text, as in Figure 8.5. In a second step, all written paraphrases can be compiled in a table in
order to systematize them and to arrive at suitable categories based on them.
In an approach to open coding based on grounded theory, codes are formed from the
outset with the intention of moving away from the data towards the goal of developing
a theory. What Strauss, Glaser, and Corbin refer to as coding refers to the theoretical
classification of the data, not simply the assignment to codes (Strauss & Corbin, 1996,
pp. 43–55). This involves two steps. First anything interesting in the text is coded,
meaning that it is given an abstract label. Then you move to the code level to group the
codes and examine the relationships between them.
Using QDA software programs to construct categories for qualitative data analysis
has many advantages compared to traditional, manual methods using pen and paper.
This is true for both methods presented, for the technique of working via summariza-
tion as well as for the more abstract and theory-oriented approach of grounded theory.
When using QDA software, you stay connected to the original data and are not forced
to page through hundreds of sheets of text searching for specific text passages. You can
easily gain an overview of how frequently certain codes, concepts, and categories appear
in the data, and you can find them, summarize them, and create categories for them
that can be used in your analysis. Likewise, you can easily find typical examples to help
define the given categories.
For documentation purposes, you can always demonstrate on which text passages a
given category is based. Moreover, you can document the different stages of building
categories. You can also determine a category’s semantic context because all of the text
passages that are linked to a given category are summarized in a list.
If you do not wish to assign codes to relevant text passages when you deal with
the text for the first time, we recommend that you work with the QDA software
programs in two phases. First, you mark important paragraphs and then code them.
For instance, MAXQDA allows you to highlight text electronically in different col-
ours. Like when highlighting on paper, you can start by simply highlighting text
passages that are important for your research question. Then you can go back
through the data a second time to assign codes to those important passages and
create categories.
Table 8.2 shows that QDA software can be integrated into every phase of thematic
text analysis to assist in the analysis very effectively, and it is by no means a complete
list of the capabilities that QDA software programs offer. Rädiker and Kuckartz (2020)
provide further detailed information on computer support for the analysis of guided
interviews; their presentation essentially follows the approach of structuring QCA.
QDA software can be particularly useful for condensing and summarizing coded
texts. Such techniques are described in detail in Chapter 5. The so-called ‘Summary
Grid’ of the software MAXQDA allows a systematic form of category-based summaries.
Figure 8.6 illustrates working with the Summary Grid. In the left-hand panel you will
find the thematic matrix, as described in Chapter 3. The columns of the matrix are
formed by the various interviews, and the rows represent the categories. Once you click
on a cell in the thematic matrix, the associated coded segments of the person and the
category are displayed in the middle pane. In the right-hand panel, the researchers can
then write a summary of the coded segments.
(Continued)
(Continued)
As described in Chapter 6, there are two different ways to evaluate cases, such as
persons, families, or institutions. Either you can decide from the start to analyse the
cases based on the categories, which requires you to read through the collection of
relevant text passages (phases 4 and 5), determine adequate characteristics based on the
definition of each evaluative category, and assign the given respondent an assessment
variable or level (e.g., sense of responsibility → high level). Or you can decide to con-
duct detailed evaluations of every relevant text passage (e.g., meaning that you evaluate
the level of responsibility expressed in each of the text passages). For this, the levels
‘high’, ‘moderate’, ‘low’, and ‘not to be classified’ can be defined as sub-categories for
the code ‘sense of responsibility’ and assigned to the appropriate text passages.
If you coded by smaller sections and defined the characteristics of the evaluative
categories as sub-categories, you will have to aggregate your codings at the case level.
You may have evaluated several relevant text passages with different values for a given
participant. For instance, some text passages show that a person feels very responsible
towards climate change, other text passages show only moderate responsibility. In some
QDA programs, these evaluations can be automatically converted into categorical vari-
ables. Then a variable with the name ‘sense of responsibility’ is created, and automatically
populated with the value of the most frequently assigned characteristic or level for each
respondent. The most frequently assigned characteristic or level may be unclear because
two or more sub-categories have the same frequency, in which case the value is set to
‘undecided’. As a result, the coders must re-examine the given text passages and assign
an appropriate value manually.
Some QDA programs allow two or more analysts to code independently. In MAXQDA,
for example, this works as follows. Two coders code independently of each other. After
all coding has been done, the software compares the codings, and calculates a coefficient
of agreement per category as well as the intercoder agreement in the form of percentage
agreement and the coefficient kappa (Kuckartz & Rädiker, 2019). A detailed table of agree-
ments per code then shows where differences exist between coders. The analysts take a
closer look at the problematic passages and try to find a consensual solution. At times,
the definitions in the coding guidelines must be modified or made more precise. If the
coders cannot reach an agreement, they can record their arguments in the form of a
memo, which can be discussed later with the research group or project management.
After the entire data set has been coded with the evaluative categories in phase 5,
QDA software offers a variety of qualitative and quantitative analysis possibilities.
As quantitative analyses, frequency analyses can first be carried out for each evaluative
category, that is, an overview is obtained of how many of the interviewed persons were clas-
sified as ‘sense of responsibility → high’ or as ‘sense of responsibility → low’. Corresponding
graphical representations can also be created such as pie charts or bar charts.
Furthermore, statistical correlation analyses between the different evaluative catego-
ries are possible. For this purpose, the matrix of the codings made can be transferred to
a statistical analysis program. Some QDA software packages, such as QDAMiner and
MAXQDA, also offer statistical functions, so that no transfer is necessary and the calcu-
lations can be made directly in the QDA software. Questions can then be answered such
as ‘What influence does a person’s sense of responsibility have?’ or ‘With which other
categories does sense of responsibility correlate?’. Since the assessment categories usually
have an ordinal scale level, rank correlation methods such as Spearman’s rho (Kuckartz
et al., 2013, pp. 216–219) are appropriate here. Correlations to socio-demographic char-
acteristics can also be examined by means of cross-tabulation analysis: Cross-tabulations
can be created, for example between sense of responsibility and level of education.
The spectrum of conceivable qualitative analyses is not much smaller than that of the
quantitative analyses. The statements (coded segments) for each value of an evaluative
category can be compiled and compared with each other in a matrix representation. For
example, for all persons with a low sense of responsibility, their definitions of the big-
gest world problems can be compiled and compared with those of persons with a high
sense of responsibility. Furthermore, the different values can serve as selection criteria
for accessing statements on other thematic categories. For example, what do persons
with a high, moderate, or low sense of responsibility say about their sources of informa-
tion on the topic of ‘learnability’? Finally, the evaluative categories can also be included
in visualizations and concept maps.
(Continued)
If you have conducted an evaluative text analysis, the appropriate assignments are already
saved as case variables. Now you can examine the individual groups formed by combina-
torics separately or compare them with each other. The characteristics saved as case
variables can be statistically analysed directly in the QDA software (if the software provides
corresponding functions) or the characteristics can be exported to statistical software in
order to create crosstabs, which can then serve as the basis for building types via the
method of reduction.
Type-building analysis benefits even more from the use of QDA software when you
rely on statistical approaches (e.g., cluster analysis, factor analysis, and correspond-
ence analysis) to help construct the types. This is particularly useful when you are
working with a large number of cases and/or with an attribute space that contains a
large number of attributes. After exporting the matrix of the attributes to be used to
construct the types into a statistical software program, you can construct natural
typologies using cluster analysis methods. The method assigns the respondents to
types, which can then be imported back into the QDA software for the next phase of
the analysis.
QDA software can also be useful for constructing types based on case summaries. In
this case, you can start with the case summaries and manually assemble them in groups
that are as homogeneous as possible according to their similarities. This can be done
without computer assistance; however, you can still rely on the computer software to
present your results, such as by creating diagrams to visualize the types and the respond-
ents who have been assigned to them.
through the texts extensively. QDA software can support such processes, which are char-
acterized by exploration and serendipity – the accidental finding of what was not
originally sought (Merton & Barber, 2004). The programs offer a toolbox full of useful
features and procedures that can each be used creatively as well as usefully combined
with others (Kuckartz & Rädiker, 2019). Describing all of the available tools and their
features goes beyond the scope of this book; thus, we will present a brief outline of those
functions that go beyond building categories and coding data that are especially useful
for qualitative content analysis.
Integrating multimedia features can also involve linking pictures, graphs, and more
to the texts. This allows you to examine people or groups as well as places, which can
increase the clarity of field research.
However, these new possibilities are accompanied by some quite serious problems.
One of these is the problem of anonymity. This has always been a problem in qualita-
tive research that should not be underestimated. Due to the new technology, this
problem has increased dramatically. In fact, it is almost impossible to make qualita-
tive data that include audio and video anonymous: voice and video recordings are
difficult to edit in such a way that a participant’s privacy is protected. Secondary
analysis of this data is also problematic, as even if respondents provide written con-
sent, researchers have to ask themselves if it is really necessary for the data to be
circulating for decades.
Visualizations
In many scientific disciplines, visualizations are a standard part of an analysis and are
used to assist in diagnosis and discovery as well as the presentation of results. Medical
or climate research would not be the same without images, and in a variety of disci-
plines statistical approaches would not be the same without suitable graphs and
diagrams of causal models. The idea of using diagrams, tables, and other visualizations
within the framework of qualitative data analysis is not new; nearly two decades ago the
proponents of grounded theory used diagrams to present their concepts (Strauss, 1991,
pp. 238–273; Strauss & Corbin, 1996, pp. 169–192), and Miles and Huberman wrote
their comprehensive book Qualitative Data Analysis: An Expanded Sourcebook as early as
1994, which presents visualization techniques in detail and is still very much worth
reading because the authors draw attention to different forms of data presentation (cf.
Kuckartz, 2010a; as well as the new edition of Miles et al., 2020).
Three of the various visualization techniques available in QDA software for present-
ing the results of QCA are described below: Visualizations that present the thematic
progression of an interview; visualizations of the categories within each interview; and
case-based concept maps.
2
The illustration was created with the help of MAXQDA’s ‘Codeline’ function.
This figure was created using the ‘Code Matrix Browser’ in MAXQDA.
3
inspire the analysis. If the word frequency review is only carried out after the qualita-
tive analysis, it also offers an interesting opportunity for comparison.
QDA software generates not only the output of frequencies for individual words
but also listings of repeatedly occurring phrases, or multi-word combinations such as
‘freedom and security’ or ‘with my family’. In so-called ‘word trees’ (Wattenberg &
Viégas, 2008) such word combinations can be displayed in their broader context. In a
sense, this is a combination of the keyword-in-context, word combinations, and word
frequencies tools in one visualization. In the example in Figure 8.10, a word tree can
be seen: In the texts analysed, the most frequent word before the word ‘family’ is the
word ‘my’. The most frequent word before ‘my family’ is ‘with’, which is therefore
placed at the top, and the second most frequent word before ‘my family’ is ‘for’,
placed in second position from the top.
Figure 8.10 Word tree with focus on the word ‘family’ (detail)
subsequent statistical analyses, but it can also simply be used to explore heuristically
and make qualitative researchers aware of specific words or word combinations and the
passages in which they appear. Dictionary-based content analysis is different from the
other text analysis approaches presented in this book because it is an automated
approach, which means that the coding is done by a computer rather than people.
Thus, some research questions can be answered very quickly. For example, we con-
ducted a dictionary-based analysis of statements that students made regarding the
proposed increases to tuition fees. We were very quickly able to identify the different
aspects and topics that the students named if they addressed social concerns or social
inequality, if they thought that the quality of their studies would be improved by
increased tuition and fees, or if they named legal aspects, such as the argument that the
German constitution guarantees every individual’s right to an education and that
imposing additional fees would be illegal. Using the word frequency lists, the terms
were assigned to categories within the dictionary and the statements were coded auto-
matically. Then it was easy to test our hypotheses, such as ‘Students connect the issues
surrounding increased tuition fees more frequently with social inequality than with
any improvements in the quality of their studies’ or ‘If a student names social aspects,
he or she also names legal aspects’.
Since word-based analyses are limited to single words or short phrases and cannot
decipher the ambiguity of words, they are of limited use when it comes to complex
research questions. However, such techniques may sometimes supplement qualitative
text analysis. They provide a different perspective on the data, draw attention to indi-
vidual words, and thus allow additional forms of analysis that can help researchers to
discover relationships within the data that may not have been visible in other ways.
Furthermore, word-based analyses are not very time- or labour-intensive because they
can be conducted on the data without further preparation.
Researchers in a team can collaborate very well using QDA software, because every
step of the work is recorded, so that it can be determined at any time, for example, who
coded which parts of the text with which categories and who wrote which paraphrase
or memo.
For all three strategies of qualitative content analysis described in this book –
structuring, evaluative, and type-building QCA – we have described in detail in this
chapter how to carry out the individual steps with QDA software. We have paid spe-
cial attention to how QDA software can support the analysis phases after the coding
of the data, and we have presented a variety of options for qualitative and quantita-
tive analytical procedures. The options we have described are intended to serve as
suggestions; there is no requirement for all of the possibilities described here to be
realized, and we do not claim that our analysis suggestions are exhaustive. There are
no limits to your creativity for further, new forms of analysis after coding.
Chapter objectives
In this chapter you will learn about:
they attempt to find a new way to determine standards rather than simply blindly reject-
ing or accepting the classical standards of quality. Their perspective encourages researchers
to reformulate quality standards that are relevant to research institutions and can be used
to analyse research proposals. Seale and Hammersley’s subtle realism (Seale, 1999b, p. 469)
serves as the basis for the following observations on the subject of quality standards within
qualitative text analysis. Their observations are based on three premises. First, the validity
of knowledge cannot be determined with certainty because assumptions can only be
judged according to plausibility and credibility. Second, phenomena exist independently
of our assumptions about them, although our assumptions can vary in closeness of fit.
Third, reality is accessible from several perspectives we may hold on the phenomena.
Research aims to present reality, not to reproduce it. In empirical qualitative research, the
main question is how well the researchers’ representations are founded in the data.
In the following, we will discuss quality standards for QCA. It seems useful to distin-
guish between internal standards of quality (i.e., authenticity and credibility) and external
standards of quality (i.e., how well you can transfer and generalize a study). The terms
internal and external standards of quality intentionally refer to the terms internal and
external validity, which stem from the classical hypothetical-deductive research para-
digm. Our terms demonstrate that the classical standards cannot simply be copied over
to qualitative research; rather, they have to be modified and extended in order to take
into account the procedural nature of qualitative research (Flick, 2020). For QCA as a
method of analysis, internal standards of quality are primarily relevant, while transfer-
ring and generalizing the results depends more on the structure of the entire study – its
design as well as the methods used for sampling. As with the classical quality standards
of internal and external validity, internal quality is considered critical for a study’s
external quality.
(a) In relation to data collection and transcription, these points are important:
• How were the data recorded? Using audio or video recordings, for example?
• Was an interview assessment document (postscript) prepared in which the
interview situation and specifics were recorded? When was the postscript prepared?
(b) The following questions are important for the implementation of the QCA:
On the last point, additional information is provided in Section 9.6. Also, a decisive
criterion for potential reviews and expert opinions is whether the methodological pro-
cedure for QCA is transparent and reflected. For reviewers, it is a considerable advantage
if you have worked with QDA software, because then it is very easy to see how elaborate
the categories are, how reliable the assignments of text passages to categories are, and
what degree of reflection is shown in the written memos.
The criteria mentioned mostly focus on the procedural aspects of the research process,
and place less emphasis on static evaluations as they are calculated in quantitative
research, for example, in the form of coefficients of intercoder reliability. Since the process
of coding is of central importance in QCA, the question of the quality of the codings and
the agreement of the coders will be considered in more detail in the following section.
Table 9.2 Coding table for ten coding units, two coders, and one category
At first glance, it can be seen that many codings match (unit 1, 2, 4, 5, 7, 8, 10),
but there are also some non-matches (unit 3, 6, 9). The table of all coding units can
be condensed into a table with four cells as shown in Table 9.3. In the main diago-
nal, the matches are shown, namely in cell a (1/1 = both coders have coded the
category) and d (0/0 = both coders have not coded the category). The non-matches
are found in cells b (1/0 = only coder 1 coded the category) and c (0/1 = only coder
2 coded the category).
Table 9.3 Table of matches for one category and two coders (general form)
Coder 1 Coder 2
Coded Not coded Total
Coded a b a+b
Not coded c d c+d
Total a+c b+d N=a+b+c+d
For the example of Table 9.2 with ten coding units, the agreement table is given in
Table 9.4. In three coding units (1, 4 and 8) there is agreement in not coding the catego-
ries (cell d), in four units the category was coded in agreement (cell a), and in a total of
three units we find non-agreement (once 0/1 in cell b; twice 1/0 in cell c).
Table 9.4 Table of matches for one category and two coders
Coder 1 Coder 2
Coded Not coded Total
Coded 4 1 5
Not coded 2 3 5
Total 6 4 10
p0 = (a+d)/N.
For the data in the example above, the result is 0.7 (i.e., 70% of all codings match).
More commonly used than this simple percentage measure of agreement is Cohen’s
kappa reliability coefficient. Kappa is based on the assumption that a certain level of
agreement would be expected even if the coders were to assign categories to the coding
units purely by chance. In our example, we proceed as follows. The corresponding prob-
abilities are calculated via the marginal frequencies. The probability that coder 1 does
not code a coding unit is
The estimated probabilities for random matches are now, for a random match in non-
coding,
The estimated total proportion pe of random match is the sum of the two probabilities:
The kappa coefficient according to Cohen (1960) incorporates this expected frequency
of random matches into the calculation of intercoder reliability. Kappa is defined as
p pe
0
.
1 pe
Kappa here is therefore (0.7 – 0.5) / (1 – 0.5) = 0.4 and thus the coefficient is consider-
ably lower than the relative or percentage agreement (p0).
This calculation of kappa can easily be extended to more than one category, for
example to four categories, resulting in Table 9.5. Cell a contains the number of coding
units for which both coders have applied category 1 (so these are agreements). Cell b
contains the number of coding units for which coder 1 has applied category 1 while
coder 2 applied category 2 (so this is a disagreement). Therefore, only the cells a, f, k,
and p on the diagonal contain the agreements; all other cells indicate disagreements.
Coder 1 Coder 2
Cat. 1 Cat. 2 Cat. 3 Cat. 4 Total
Cat. 1 a b c d a+b+c+d
Cat. 2 e f g h e+f+g+h
Cat. 3 i j k l i+j+k+l
Cat. 4 m n o p m+n+o+p
Total a+e+i+m b+f+j+n c+g+k+o d+h+l+p N
The probabilities for each of the four categories are now calculated for both coders
via the marginal frequencies and then the estimated probabilities of random agreement
are determined for all four categories, summed and entered into the kappa formula.
So how is the level of the kappa coefficient evaluated? What is interpreted a good or
very good value for intercoder reliability? As a rule of thumb, kappa values above 0.6
are assessed as good, above 0.8 as very good. It should be noted that the level of kappa
is strongly dependent on the distribution of the marginal frequencies, and kappa values
can be unreasonably small in the case of asymmetrical distributions (Feinstein &
Cicchetti, 1990). There are alternative measures of intercoder reliability such as
Krippendorff’s alpha and Scott’s pi (Krippendorff, 2004b; Zhao et al., 2012) which,
however, are beyond the scope of this chapter.
A very important difference is that in qualitative content analysis the data is usually not
segmented in advance. The processes of segmenting and coding belong together – they form
a unity. In most situations in qualitative content analysis, units of meaning are coded, that is, the
coders are free to determine the beginning and end of such a meaning unit.
This means that a table like Table 9.2 cannot be created easily, because the coders may
have segmented differently. And yet this table is the starting point for all subsequent
calculations. Before discussing ways and means of calculating coefficients of agreement
despite these difficulties, however, it should first be noted that there are two ways of
ensuring the agreement of coders in QCA: a qualitative way via the collaborative checking
of codings (‘consensual coding’) and a quantitative way via the calculation of percentage
agreement and, if necessary, also of a suitable coefficient that takes the probability of
chance agreement into account.
comments on the coding in question if QDA software is used. Next follows the joint
part, which is much more effective if QDA software is used to visualize the codings of
both persons simultaneously on the text. Differences in code applications are dis-
cussed and the coders look up the corresponding category definitions as their
guideline. Where the differences are resolved with mutual agreement, this coding is
recorded; possibly a suggestion for improving the category definitions is also noted,
which can then be brought to the next team meeting.
What happens if the two coders still cannot agree? In this situation, a third person can
be called in. It is best to appoint a person in the team beforehand who will take on the role
of supervisor for the coding phase and then decide in each instance, after listening to the
arguments, which coding should take place. With very important disagreements, the deci-
sion can be postponed to a team meeting. However, this should only be done in the case
of differences of fundamental importance for the development of the category system.
What pitfalls should one expect? Consensual coding is laborious and requires a will-
ingness to put forward good arguments for the codes applied. It is therefore not
expedient or desirable to act according to the principle of ‘the wiser head gives in’.
Routine, as would be established by fixed coding pairs, should therefore be avoided.
Table 9.6 Comparison of coding for two coders and multiple categories
Of course, this extension of the coding unit to the entire case is associated with
considerable loss of information. In contrast to the very fine-grained coding of seg-
ments, only the entire interview is considered here – highly aggregated. The loss of
information by defining documents as coding units is reduced not only if the coders’
comparison table takes into account whether the category was coded, but also when the
frequency of coding of the category in the document is taken as a basis. The compari-
son table in Table 9.7 serves as an example. For this table, too, there is agreement in 4
out of 5 categories (i.e., 80%).
Table 9.7 Comparison of frequency of coding for two coders and several categories
The most difficult way to measure the level of coder agreement in QCA is to make a
segment-specific calculation. Among others, the following situations can be distinguished:
1 Coder 1 and coder 2 both coded exactly the same text passage with the same
category.
2 Coder 1 and coder 2 both coded the same text passage with the same category,
but not with exactly the same segment boundaries (i.e., the codings overlap but
are not identical).
3 Coder 1 has coded the text passage with category A, while coder 2 has coded
exactly the same text passage with category B.
4 Coder 1 coded the text passage with category A, while coder 2 coded an
overlapping passage with category B.
5 Coder 1 has coded the text passage with category A, while coder 2 has not.1
Before one can create a category-specific matching table similar to the one in Table 9.3
it is first necessary to clarify how a coding unit is to be defined and what is counted as a
match. Let us consider the above five situations:
• Situation 1 and situation 3 are not problematic: the coding units are clearly
defined (‘the same text passage’). In situation 1 we have a match, in situation 3
1
Not yet considered here are the complex cases where, for example, coder 1 has coded a
segment with category A and coder 2 has coded the same segment with categories B, C,
and D.
Coder 1 Coder 2
Coded Not coded Total
Coded a=6 b=4 a + b = 10
Not coded c=3 d=0 c+d=3
Total a+c=9 b+d=4 N = 13
But what about the calculation of a chance-corrected coefficient, namely kappa? Due
to the method of qualitative coding, which normally takes place without a priori defini-
tion of coding units, the cell d always remains zero, because segmenting and coding are
one and the same process and so logically there cannot exist any segments that were
not coded by both coders, because, for a segment to exist, it must have been coded by
someone. The calculation of the random agreement via the marginal totals is therefore
not possible. Instead one can follow a suggestion by Brennan and Prediger (1981): the
random agreement is calculated based on the number of different categories from which
the two coders could choose. Supposing the category system consists of 10 categories.
Then the expected value of random agreement is
1
pe = = 0.10,
Number of categories
while kappa is
Usually, the number of categories and sub-categories in QCA is greater than 10, so that
kappa is usually only slightly lower than the percentage agreement. However, it must be
asked what ‘chance-corrected’ means in qualitative coding. With fixed coding units and
mutually exclusive categories, it seems reasonable to assume that, for example, with 10
categories, the probability of randomly coding the ‘right’ one is equal to 1/10. With free
segmenting and coding, one could accordingly calculate the random correction based
on words. For example, given a text with 3000 words, how likely is it that one particular
word or phrase will be randomly coded with the same category by two independent
coders? That probability should asymptotically tend to zero as the size of text increases.
It can be concluded that the calculation of kappa (or any other chance-corrected coeffi-
cient) only makes sense if coding units are defined a priori.
checking for agreement in the early stages of coding, the calculation of coefficients,
especially the detailed calculation per category, can save a lot of time that can be well
used elsewhere in the project. The percentage agreement figures can easily be used to
identify the categories that are problematical and need to be worked on together; the
category definitions can be inspected again and revised if necessary.
In this way, the calculation of intercoder agreements, if done with QDA software, can
be an effective support in reaching a high level of agreement quickly and effectively in
the team. It also becomes clear which coders deviate particularly frequently. Another
advantage is that an overall coefficient is calculated, which might be a helpful indicator
of the quality of coding.
In addition, the calculation of the coefficients may have a legitimizing function.
Reviewers who come from the field of quantitative research and are committed to the
hypothetico-deductive paradigm often request that information on the intercoder reli-
ability be provided. A lack of such information is seen as a deficit. Similarly, preference
is often given to the kappa coefficient over the calculation of percentage agreement. If
kappa is not calculated, this is often considered a weakness, although the random effect
in an analysis with a typical qualitative category system is usually very small. In qualita-
tive coding with free segmenting and coding, the calculation of kappa makes little sense
because here the model on which kappa is based is simply not appropriate for the real-
ity of the coding process in QCA. If reviewers nevertheless require the calculation of
kappa, then coding units should be defined a priori.
• Discussion with experts (peer debriefing). This means regular meetings with
competent persons outside the research team. These experts comment on the
approach and the first results of the project and, if necessary, draw attention to
phenomena and facts that are easily overlooked.
• Discussion with research participants (member checking). This means the discussion of
the analysis results with the research participants themselves in order to receive
informed feedback on the research results in the sense of ‘communicative validation’.
• Extended stay in the field. A longer stay in the field or a return to the field can also
help to avoid hasty diagnoses and false conclusions in the analysis of the material.
• Triangulation or use of mixed methods. Techniques of triangulation and the
combination of different research methods (Creswell & Plano Clark, 2018; Denzin,
1978; Flick, 2018b; Kelle, 2008; Kuckartz, 2014b) result in more diverse perspectives
on the research results and thus also the possibility of increasing generalizability.
what your real motives are for extending the results. Why do you want to generalize at
all? Often the answer is that it will not be enough for most researchers to only make
statements about a small group of perhaps only 12 respondents after a relatively elabo-
rate study. Such a minimalist claim would probably also disappoint the scientific
community to whom the results are presented. But it is not only the problem of the
small number of research participants that plays a role here. Even with 100 or 200
respondents, which is a large number of cases for a qualitative study, questions about
the composition of the sample and the selection process are still relevant. For example,
if I want to find out about the consequences of unemployment, then the selection of
100 respondents in a single small town in Saxony may not be adequate to answer far-
reaching questions about the chances of success for political measures for the
unemployed. The implicit claim by such a researcher is that the selected small town
should be the laboratory for many towns, or at least for many small towns. Let us look
at what generalization strategies are available in social science research:
2
“In Germany, both parents can receive a parental allowance for a maximum of 12 months.
This can be extended by 2 months with the help of the so-called ‘fathers’ months’ or
‘fathers’ time’.
It should be obvious that in qualitative studies, generalization strategies 1 and 2 are usu-
ally not available. Due to time and budget constraints, the third strategy is also rarely
conceivable. So, if the classical critical-rationalist strategies with random sampling and
inferential statistics are not applicable, then the argumentative justification of general-
izations is required, that is, a new theory must be developed and it must be convinc-
ingly demonstrated that it can explain the research results as a ‘case of’. In doing so,
one should be aware of the risky nature of abductive inferences – as Peirce (1931/1974,
p. 106) puts it: ‘Abduction merely suggests that something may be’.
Because of this general problem, it seems to make sense to give preference to the term
‘transferability’ over ‘generalization’, because transferability has a certain inherent limi-
tation of expansion, while generalization suggests boundlessness. The goal of
generalization in the sense of critical-rationalist strategies is – as is well known in
qualitative research – not achievable anyway. Furthermore, according to the self-under-
standing of qualitative research, this goal does not make sense at all, because of what is
known about difference and diversity, about the context- and situation-bound nature of
phenomena in the social fields. For example, there are working-class young people who
study at universities, although all research shows that they are at a great disadvantage.
This social fact is true even if it only applies to 5% of a group. Generalized statements
are not the primary goal of qualitative research and probabilistic ones (e.g., ‘The
chances of a working-class child getting a degree are four times less than those of a
middle-class child’) are a matter for quantitatively oriented social research. When the
researchers of the famous Marienthal study wanted to know how large was the propor-
tion of the different types of attitudes they had identified, they did not rely on the
numbers of their qualitative study but conducted a survey to determine the exact distri-
bution of the identified types of attitudes. In the context of qualitative research,
percentages and distributions can only be indications or rough estimates. So, beware of
such generalizing statements as ‘This is how doctors think’, ‘This is how nurses or teach-
ers think’, or even ‘This is how Americans think’. If, however, random sampling was
carried out within the framework of a qualitative study and the number of cases is suf-
ficient, confidence intervals and parameters can of course also be calculated, but such
studies are very rare.
Note that the generalization strategies common in quantitative research cannot be
used, or can only be used very rarely, in qualitative research due to its sampling strate-
gies and limited sample sizes. The question of transferability to external contexts should,
however, be considered in the results report and, according to Flick (2009, p. 276), con-
crete steps should be taken to check and assess transferability.
particularly throughout the data analysis process. Doing so enables you to accumu-
late a good deal of material that you can use to write your final research report,
which simply constitutes the last stage of this continuous writing process.
At the end of the research work there must be results – as was quoted at the begin-
ning of this book in the forum contribution of a diploma student: ‘One wants to report
results, after all’. When integrating the various fragments of content that have emerged
during the analysis, one should always ask oneself: What is my research question?
Whatever you include in the report should answer this question, as it indicates how
relevant and how useful the information is in practice or for further research.
Everything that you have written in the course of the analysis forms the basis for the
research report, including:
Thus, when you start writing the research report, you have a sort of inventory of
everything that you have already produced over the course of the research process.
If you are working in a team, gaining an overview of this inventory may prove to
be a time-consuming task; however, it will also reveal any gaps or areas that require
additional preparation. A good amount of literature has been published that provides
instructions for writing research papers; thus, we will not list them here. As various
authors correctly point out, not all researchers rely on the same writing process or
structure. However, you should start with an outline, which could be based on the
following general structure:
Other important differentiations arise on their own. For example, the methods used to
collect data, the type and rules for transcription, and the steps in the qualitative text
analysis process will be described in your methods chapter. You can emphasize different
elements of your project depending on whether you are writing an academic thesis, a
research project funded by a third party, or an evaluation. Naturally, you may have to
adhere to certain forms more rigidly and fulfil more specific and detailed requirements
in the methods sections of academic theses than in other types of papers. In evaluations,
the results, the assessments carried out by the evaluators and their consequences are
usually of primary importance.
When writing the research report in qualitative research, researchers often encounter
a phenomenon that Huberman and Miles refer to as ‘data overload’. When you have
collected so much interesting data, it can be difficult to see the wood for the trees, so to
speak. Thus, it can be difficult to select the results and their underlying data. What
should you report and what should you omit? Why should you include this case sum-
mary and not that one? Why are you focusing on a given category?
Unfortunately, researchers often use up most of their time and energy on transcrib-
ing and coding the data. These first few steps of the analysis process can understandably
consume large amounts of time, which means that researchers lack the time and energy
necessary to conduct complex analyses and compose their research reports. We recom-
mend keeping the entire research process in mind, allotting sufficient time for writing
and recording your results after you have completed the analysis, and – as mentioned
above – thinking about what you would like to write throughout the entire analysis.
While writing, you might worry that your results could have repercussions on the
field that you have researched. It is essential that you anticipate these potential effects
and include them in your report. This is especially true for evaluations. The Standards for
Evaluations, originally developed by the Joint Committee on Standards for Educational
Evaluation (JCSEE; see Yarbrough et al., 2011), emphasize the need for fairness:
You should take such standards into account when writing a report and consult the cli-
ent and stakeholders before finalizing your text, if necessary.
3
See the program evaluation standards on the website of the American Evaluation
Association at https://www.eval.org/About/Competencies-Standards/Program-Evaluation-
Standards (accessed 20 March 2022).
Qualitative researchers feel the same way about verbal data, which should show the scien-
tific community what the result of the analytical work looks like. Wanting to quote
passages from the open interviews, for example, is perfectly natural and there is no reason
not to include quotes in your research report. Every quote must be labelled as such and
omissions must be noted. Like quotations from other sources, all quotes should contain
information regarding their source, including the interview name and paragraph or line
number. For example, ‘(R07: para. 14)’ and ‘(Ms. Stone: lines 311–315)’ represent valid
citations with source information. The first example includes an abbreviated interview
name and a paragraph number, and the second example contains an interviewee pseudo-
nym and line numbers.
Quotations should be used sparingly; they should not comprise more than a quarter
of the results section, even in a thesis. Reproducing authentic ‘sound bites’ may seem
attractive, but it gives the scholarly paper or thesis a non-analytic character, which
should be avoided.
Be aware of the danger of selective plausibility – using original quotations to justify
each analytical finding. It can be tempting to do so, but it will likely make your readers
suspicious. Thus, you should present contrasting statements in the research report and
include a broad spectrum of answers using quotations.
Documentation
Transparency and auditability were mentioned above in Section 9.1 as special quality
criteria of qualitative research. This means that good documentation should be pre-
pared. What should you document and in what form, for example, in an academic
thesis? What information must be kept confidential? What should your evaluators be
able to understand and verify, if desired?
In the methods section of your research report, you should describe your chosen
method of QCA clearly and comprehensibly:
• How were the data selected for the analysis? How voluminous was the material?
What was the nature of the material?
• Which type of QCA did you use and with what aim?
• How did you explore the data?
• How was the category system developed? Did it consist primarily of deductive or
inductive categories, and how did it change over the course of the analysis, and
why? What types of categories were used? If inductive categories were formed,
how much data was used for this?
• How was coding organized? What criteria were used to define the segment
boundaries? How many people worked on the coding and how was the joint
coding cycle organized, if applicable?
• What procedures were used to ensure and verify coding quality? Was coder
agreement assessed? With what results? How were non-agreements dealt with?
• How were the coded data analysed? What role did case-oriented and category-
oriented approaches play?
• Which QDA software was used and which functions facilitated the analysis?
At a minimum, the categories that were central to the analysis should be presented in
the text itself. However, the following elements also should be included in the appendix
of a thesis or research report:
• Written documents that are important to the study, such as covering letters
• Transcription rules and references to relevant standards (which can also be
included directly in the text)
• The interview guide (if applicable)
• The accompanying questionnaire (if you have used one)
• Information regarding the length of the individual interviews or at least the
average duration and the range of the lengths of the interviews
• The code book, that is, the documentation of the category system, including
examples
• One or more transcripts that serve as examples of the data collected and the type
of transcription, if requested by the evaluators.
Moreover, you should submit the following data in electronic form (if the evaluators
ask for it):
• The final version of the project file, if QDA software was used in the analysis
• Transcripts of the anonymized original data in a conventional standard format
(DOCX, RTF or PDF); this is not necessary when QDA software was used, as the
project file contains the transcripts.
4
https://www.ukdataservice.ac.uk/deposit-data.aspx (accessed 20 March 2022).
to reduced possibilities for standardization. Laudel and Bielick (2019) report numerous
practical research problems in archiving guided interviews, and Corti et al. (2005) dis-
cuss the potential benefits and problems of secondary analyses of qualitative data in
their introduction to a special issue of the online journal FQS – Forum Qualitative Social
Research (www.qualitative-research.net).
The following particular aspects should be taken into account when considering the
secondary use of data collected and analysed with QCA: first, the informed consent of all
respondents to storage and secondary use of their data; and second, the sufficient
anonymization of the data. On the Qualiservice website of the Research Data Centre for
Qualitative Social Science Data at the University of Bremen, Germany, you can find numer-
ous helpful tips and handouts, including template forms for consent declarations.5
The preparation of research data for subsequent use is supported by QDA software
with various functions:
• The original data can be exported (without the coding done by the analysts),
whereby some software packages also allow the automatic masking of marked areas.
• Individual QDA packages offer special functions for archiving and exporting
the documents analysed, optionally with the associated audio and video files, the
standardized background information (e.g., socio-demographic data), and the
category system used.
• Entire projects or even just the code book can be stored in a general exchange
format for QDA programs, the so-called REFI-QDA standard (www.qdasoftware.
org), which is considered the preferred data format by some data archives.
and well-documented category system and trustworthy coding of the data. If the data
are coded by several analysts, we recommend the procedure of consensual coding. If
researchers are more oriented towards the classical quality criteria of reliability and
validity, coefficients for intercoder agreement can also be calculated. We describe in
detail how this can be done.
At the end of a project there is always a written product, perhaps a research report, a
master’s thesis, or a dissertation. It is often essential to defend one’s work in front of a
sceptical audience. Research always takes place under specific circumstances and restric-
tions. For example, researchers do not have enough time, or do not have enough
financial resources, field access was not as expected, and many other unexpected cir-
cumstances may arise. There is much written in textbooks on methods and methodology
that cannot always be realized satisfactorily in research practice. Our advice is: work
carefully and trust your work. Remember that the most important thing is to achieve
useful results – ultimately to change the world with your research and, if possible, to
improve it. So do not be upset when you hear criticism from a sceptical audience.
Sometimes the criticism is levelled at researchers who work with the method of qualita-
tive content analysis that QCA has no background theory. Politely counter such
criticism by saying that you see QCA not as a world-view or methodology but as a
method, and that what matters to you is producing substantive results. Be proud of the
grounding that your results have in the empirical data, in their empirical justification.
In the end, it is the findings that matter. Methods can provide valuable support, but
they are not an end in themselves.
As a method of analysis for qualitative data, qualitative content analysis is meeting with
steadily growing interest and is being chosen by more and more qualitative researchers.
The method is used for research in many academic disciplines and fields of practice,
including education, sociology, political science, psychology, ethnology, social work
and social pedagogy, nursing, and health science. The popularity in so many fields is
understandable, because the different variants of QCA have numerous strengths. When
compared with other methods for qualitative data analysis, QCA, as presented in this
book, has many advantages:
If we take Kracauer’s (1952) demand for codification from the beginnings of QCA as a
yardstick, there has undoubtedly been great progress along this path in recent years.
In some points there is still a need for further development, for example regarding the
development of standards. Time does not stand still, and so empirical research and its
methods are also subject to constant pressure to innovate. In recent years, it is above all
the social media, which have grown considerably in importance, for which new forms
of analysis and systematic methods are needed. This is certainly a field for the further
development of QCA. Social media data differ considerably from the usual qualita-
tive data such as that collected in the past, mainly by means of open interviews, focus
groups, or field research. When analysing data from Twitter, it can easily happen that
several thousand tweets are to be analysed, each of which is relatively short. Here, too,
the case-oriented perspective we have emphasized in many places in this book can play
a role: What kind of people are these, for example, who arouse widespread anger and
attract attention by insults and abuse of the worst kind? However, some maxims of QCA
can no longer be implemented when analysing data of this kind, such as the rule of
gaining as much knowledge of the data as possible in the initiating phase of the analysis.
No one can be expected to read 15,000 tweets. These new types of data demand method-
ological innovations, whereby a combination of human analytical understanding and
algorithmic power is increasingly becoming a desideratum.
It may be surprising to read at the end of this book about a future turn towards analy-
sis techniques that are based on machine learning and artificial intelligence. Further
developments and methodological innovations will drive qualitative content analysis
forward, and we believe AI will be a driving force in this. So we end, then, with the proph-
ecy of a best-selling 1960s futurology book The Year 2000: A Framework for Speculation on
the Next Thirty-Three Years (Kahn & Wiener, 1967): ‘You will experience it.’
Creswell, J. W., & Creswell Báez, J. (2021). 30 essential skills for the qualitative researcher
(2nd ed.). SAGE.
Creswell, J. W., & Plano Clark, V. L. (2011). Designing and conducting mixed methods
research (2nd ed.). SAGE.
Creswell, J. W., & Plano Clark, V. L. (2018). Designing and conducting mixed methods
research (3rd ed.). SAGE.
Creswell, J. W., & Poth, C. N. (2018). Qualitative inquiry & research design: Choosing among
five approaches (4th ed.). SAGE.
Denzin, N. K. (1978). The research act: A theoretical introduction to sociological methods
(2nd ed.). McGraw-Hill.
Denzin, N. K., & Lincoln, Y. S. (Eds.). (2018). The SAGE handbook of qualitative research
(5th ed.). SAGE.
Devi Prasad, B. (2019). Qualitative content analysis: Why is it still a path less taken?
Forum Qualitative Sozialforschung / Forum: Qualitative Social Research, 20(3), Art. 36.
https://doi.org/10.17169/FQS-20.3.3392
Dey, I. (1993). Qualitative data analysis: A user-friendly guide for social scientists. Routledge.
Diekmann, A. (2007). Empirische Sozialforschung: Grundlagen, Methoden, Anwendungen (4th
ed.). Rowohlt Taschenbuch Verlag.
Dietrich, M., & Mey, G. (2019). Visuelle Jugendkulturforschung: Trends und
Entwicklungen. Diskurs Kindheits- und Jugendforschung / Discourse Journal of Childhood
and Adolescence Research, 14(3), 293–307. https://doi.org/10.3224/diskurs.v14i3.04
Dilthey, W. (1977). Ideas concerning a descriptive and analytic psychology (1894). In
W. Dilthey, Descriptive psychology and historical understanding (pp. 21–120). Martinus
Nijhoff. https://doi.org/10.1007/978-94-009-9658-8_2 (Original work published 1894.)
Dittmar, N. (2009). Transkription. Ein Leitfaden mit Aufgaben für Studenten (3rd ed.). VS
Verlag für Sozialwissenschaften.
Dresing, T., Pehl, T., & Schmieder, C. (2015). Manual (on) transcription: Transcription conventions,
software guides and practical hints for qualitative researchers (3rd ed.). Self-published. https://
www.audiotranskription.de/wp-content/uploads/2020/11/manual-on-transcription.pdf
Döring, N. & Bortz, J. (2016). Forschungsmethoden und Evaluation in den Sozial- und
Humanwissenschaften (5th ed.). Springer.
Elo, S., Kääriäinen, M., Kanste, O., Pölkki, T., Utriainen, K., & Kyngäs, H. (2014).
Qualitative content analysis: A focus on trustworthiness. SAGE Open, 4(1), 1–10.
https://doi.org/10.1177/2158244014522633
Esser, H. (2001). Soziologie – Spezielle Grundlagen, Band 6: Sinn und Kultur. Campus.
Feinstein, A. R., & Cicchetti, D. V. (1990). High agreement but low Kappa: I. The
problems of two paradoxes. Journal of Clinical Epidemiology, 43(6), 543–549.
https://doi.org/10.1016/0895-4356(90)90158-L
Fenzl, T., & Mayring, P. (2017). QCAmap: Eine interaktive Webapplikation für
Qualitative Inhaltsanalyse. Zeitschrift für Soziologie der Erziehung und Sozialisation,
37(3), 333–340.
Festinger, L. (1957). A theory of cognitive dissonance. Stanford University Press.
Fielding, J. (2008). Double whammy? Are the most at risk the least aware? A study
of environmental justice and awareness of flood risk in England and Wales.
Gugushvili, T., & Salukvadze, G. (2021). Using MAXQDA for analyzing documents: An
example of prioritization research design in urban development. In M. C. Gizzi &
S. Rädiker (Eds.), The practice of qualitative data analysis: Research examples using
MAXQDA (pp. 107–120). MAXQDA Press.
Hammersley, M. (1992). What’s wrong with ethnography? Routledge.
Helfferich, C. (2011). Die Qualität qualitativer Daten: Manual für die Durchführung
qualitativer Interviews. VS Verlag für Sozialwissenschaften. https://doi.
org/10.1007/978-3-531-92076-4
Hempel, C. G., & Oppenheim, P. (1936). Der Typusbegriff im Lichte der neuen Logik:
Wissenschaftstheoretische Untersuchungen zur Konstitutionsforschung und Psychologie. Sijthoff.
Höld, R. (2009). Zur Transkription von Audiodaten. In R. Buber & H. H. Holzmüller
(Eds.), Qualitative Marktforschung (pp. 655–668). Gabler. https://doi.org/10.1007/978-
3-8349-9441-7_41
Hopf, C. (2016). Schriften zu Methodologie und Methoden qualitativer Sozialforschung (W. Hopf
& U. Kuckartz, Eds.). Springer VS.
Hopf, C., Rieker, P., Sanden-Marcus, M., & Schmidt, C. (1995). Familie und
Rechtsextremismus. Familiale Sozialisation und rechtsextreme Orientierungen junger
Männer. Juventa.
Hopf, C., & Schmidt, C. (1993). Zum Verhältnis von innerfamilialen sozialen Erfahrungen,
Persönlichkeitsentwicklung und politischen Orientierungen: Dokumentation und
Erörterung des methodischen Vorgehens in einer Studie zu diesem Thema. Institut für
Sozialwissenschaften der Universität Hildesheim. https://nbn-resolving.org/
urn:nbn:de:0168-ssoar-456148
Hsieh, H.-F., & Shannon, S. E. (2005). Three approaches to qualitative content analysis.
Qualitative Health Research, 15(9), 1277–1288. https://doi.org/10.1177/1049732305276687
Jackson, K., & Bazeley, P. (2019). Qualitative data analysis with NVivo (3rd ed.). SAGE.
Jahoda, M., Lazarsfeld, P. F., & Zeisel, H. (2002). Marienthal: The sociography of an
unemployed community. Transaction Publishers. (Original work published 1933.)
Janssen, M., Stamann, C., Schreier, M., Whittal, A., & Dahl, T. (Eds.). (2019). FQS special
issue ‘Qualitative Content Analysis I’. Forum Qualitative Sozialforschung / Forum:
Qualitative Social Research, 20(3). http://www.qualitative-research.net/index.php/fqs/
issue/view/65
Johnson, B., & Christensen, L. B. (2020). Educational research: Quantitative, qualitative,
and mixed approaches (7th ed.). SAGE.
Kahn, H., & Wiener, A. J. (1967). The year 2000: A framework for speculation on the next
thirty-three years. Macmillan.
Kan, M. P. H., & Fabrigar, L. R. (2017). Theory of planned behavior. In V. Zeigler-Hill &
T. K. Shackelford (Eds.), Encyclopedia of personality and individual differences (pp. 1–8).
Springer. https://doi.org/10.1007/978-3-319-28099-8_1191-1
Kelle, U. (2007). Theoretisches Vorwissen und Kategorienbildung in der ‘Grounded
Theory’. In U. Kuckartz, H. Grunenberg, & T. Dresing (Eds.), Qualitative Datenanalyse:
Computergestützt (pp. 32–49). VS Verlag für Sozialwissenschaften.
Kelle, U. (2008). Die Integration qualitativer und quantitativer Methoden in der empirischen
Sozialforschung (2nd ed.). VS Verlag für Sozialwissenschaften. https://doi.
org/10.1007/978-3-531-91174-8
Kelle, U., & Kluge, S. (2010). Vom Einzelfall zum Typus: Fallvergleich und Fallkontrastierung
in der Qualitativen Sozialforschung (2nd ed.). VS Verlag für Sozialwissenschaften.
Keller, R. (2011). Wissenssoziologische Diskursanalyse: Grundlegung eines
Forschungsprogramms (3rd ed.). VS Verlag für Sozialwissenschaften.
Kirk, J., & Miller, M. L. (1986). Reliability and validity in qualitative research. Sage Publications.
Klafki, W. (2001). Hermeneutische Verfahren in der Erziehungswissenschaft. In
C. Rittelmeyer & M. Parmentier (Eds.), Einführung in die pädagogische Hermeneutik. Mit
einem Beitrag von Wolfgang Klafki (pp. 125–148). Wissenschaftliche Buchgesellschaft.
(Original work published 1971.)
Kluge, S. (1999). Empirisch begründete Typenbildung: Zur Konstruktion von Typen und
Typologien in der qualitativen Sozialforschung. VS Verlag für Sozialwissenschaften.
https://doi.org/10.1007/978-3-322-97436-5
Kowal, S., & O’Connell, D. C. (2014). Transcription as a crucial step of data analysis. In
U. Flick, The SAGE handbook of qualitative data analysis (pp. 64–78). SAGE. https://doi.
org/10.4135/9781446282243.n5
Kracauer, S. (1952). The challenge of qualitative content analysis. Public Opinion
Quarterly, 16(4, Special Issue on International Communications Research), 631–
642. https://doi.org/10.1086/266427
Krippendorff, K. (1980). Content analysis: An introduction to its methodology. SAGE.
Krippendorff, K. (2004a). Content analysis: An introduction to its methodology (2nd ed.).
SAGE.
Krippendorff, K. (2004b). Reliability in content analysis: Some common
misconceptions and recommendations. Human Communication Research, 30(3),
411–433. https://doi.org/10.1111/J.1468-2958.2004.TB00738.X
Krippendorff, K. (2018). Content analysis: An introduction to its methodology (4th ed.). SAGE.
Kriz, J., & Lisch, R. (1988). Methoden-Lexikon. PVU.
Kuckartz, U. (1991). Ideal types or empirical types: The case of Max Weber’s empirical
research. Bulletin of Sociological Methodology/Bulletin de Méthodologie Sociologique, 32(1),
44–53. https://doi.org/10.1177/075910639103200103
Kuckartz, U. (2006). Zwischen Singularität und Allgemeingültigkeit: Typenbildung
als qualitative Strategie der Verallgemeinerung. In K.-S. Rehberg (Ed.), Soziale
Ungleichheit, kulturelle Unterschiede: Verhandlungen des 32. Kongresses der Deutschen
Gesellschaft für Soziologie in München (pp. 4047–4056). Campus. https://
nbn-resolving.org/urn:nbn:de:0168-ssoar-142318
Kuckartz, U. (2009). Methodenkombination. In B. Westle (Ed.), Methoden der
Politikwissenschaft (pp. 352–263). Nomos.
Kuckartz, U. (2010a). Einführung in die computergestützte Analyse qualitativer Daten
(3rd ed.). VS Verlag für Sozialwissenschaften.
Kuckartz, U. (2010b). Nicht hier, nicht jetzt, nicht ich – Über die symbolische
Bearbeitung eines ernsten Problems. In H. Welzer, H.-G. Soeffner, & D. Giesecke
(Eds.), Klimakulturen: Soziale Wirklichkeiten im Klimawandel (pp. 144–160). Campus.
Kuckartz, U. (2014a). Qualitative text analysis: A guide to methods, practice & using software.
SAGE.
Kuckartz, U. (2014b). Mixed methods: Methodologie, Forschungsdesigns und Analyseverfahren.
Springer VS.
Kuckartz, U., Dresing, T., Rädiker, S., & Stefer, C. (2008). Qualitative Evaluation: Der
Einstieg in die Praxis (2nd ed.). VS Verlag für Sozialwissenschaften.
Kuckartz, U., & Rädiker, S. (2019). Analyzing qualitative data with MAXQDA: Text, audio,
and video. Springer Nature Switzerland. https://doi.org/10.1007/978-3-030-15671-8
Kuckartz, U., & Rädiker, S. (2021). Using MAXQDA for mixed methods research. In
R. B. Johnson & A. J. Onwuegbuzie (Eds.), The Routledge reviewer’s guide to mixed
methods analysis (pp. 305–318). Routledge. https://doi.org/10.4324/9780203729434-26
Kuckartz, U., & Rädiker, S. (2022). Qualitative Inhaltsanalyse. Methoden, Praxis,
Computerunterstützung (5th ed.). Beltz Juventa.
Kuckartz, U., Rädiker, S., Ebert, T., & Schehl, J. (2013). Statistik: Eine verständliche
Einführung (2nd ed.). VS Verlag für Sozialwissenschaften.
Lamnek, S., & Krell, C. (2016). Qualitative Sozialforschung (6th ed.). Beltz.
Laudel, G., & Bielick, J. (2019). Forschungspraktische Probleme bei der Archivierung von
leitfadengestützten Interviews. Forum Qualitative Sozialforschung / Forum: Qualitative
Social Research, 20(2), Art. 10. https://doi.org/10.17169/fqs-20.2.3077
Lazarsfeld, P. F. (1972). Qualitative analysis. Historical and critical essays. Allyn and Bacon.
Legewie, H., & Schervier-Legewie, B. (2004). Anselm Strauss: Research is hard work, it’s
always a bit suffering. Therefore, on the other side research should be fun. Forum
Qualitative Sozialforschung / Forum: Qualitative Social Research, 5(3), Art. 22. https://
doi.org/10.17169/FQS-5.3.562
Maksutova, A. (2021). Using MAXQDA’s summary features: Developing social types in
migrant integration studies. In M. C. Gizzi & S. Rädiker (Eds.), The practice of
qualitative data analysis: Research examples using MAXQDA (pp. 135–147). MAXQDA
Press. https://doi.org/10.36192/978-3-948768058
Marshall, C., & Rossman, G. B. (2011). Designing qualitative research (5th ed.). SAGE.
Mayring, P. (1983). Qualitative Inhaltsanalyse: Grundlagen und Techniken. Beltz.
Mayring, P. (2002). Einführung in die qualitative Sozialforschung (5th ed.). Beltz.
Mayring, P. (2014). Qualitative content analysis: Theoretical foundation, basic procedures
and software solution. http://nbn-resolving.de/urn:nbn:de:0168-ssoar-395173
Mayring, P. (2015). Qualitative Inhaltsanalyse: Grundlagen und Techniken (12th ed.). Beltz.
Mayring, P. (2019). Qualitative Inhaltsanalyse – Abgrenzungen, Spielarten,
Weiterentwicklungen. Forum Qualitative Sozialforschung / Forum: Qualitative Social
Research, 20(3), Art. 16. https://doi.org/10.17169/fqs-20.3.3343
Mayring, P. (2020). Qualitative Inhaltsanalyse. In G. Mey & K. Mruck (Eds.), Handbuch
Qualitative Forschung in der Psychologie (pp. 495–511). Springer Fachmedien. https://
doi.org/10.1007/978-3-658-26887-9_52
Mayring, P. (2021). Qualitative content analysis: A step-by-step guide. SAGE.
Mayring, P., & Gläser-Zikuda, M. (Eds.). (2005). Die Praxis der qualitativen Inhaltsanalyse.
Beltz.
Medjedović, I. (2014). Qualitative Sekundäranalyse: Zum Potenzial einer neuen
Forschungsstrategie in der empirischen Sozialforschung. Springer VS.
Medjedović, I., & Witzel, A. (2010). Wiederverwendung qualitativer Daten: Archivierung und
Sekundärnutzung qualitativer Interviewtranskripte. VS Verlag für Sozialwissenschaften.
https://doi.org/10.1007/978-3-531-92403-8
Merten, K. (1995). Inhaltsanalyse: Einführung in Theorie, Methode und Praxis (2nd ed.).
Springer Fachmedien.
Mertens, D. M. (2018). Mixed methods design in evaluation. SAGE.
Merton, R. K., & Barber, E. (2004). The travels and adventures of serendipity: A study in
sociological semantics and the sociology of science. Princeton University Press.
Mey, G., & Mruck, K. (2011). Grounded Theory Reader (2nd ed.). VS Verlag für
Sozialwissenschaften.
Miles, M. B., & Huberman, A. M. (1984). Qualitative data analysis: A sourcebook of new
methods. SAGE.
Miles, M. B., & Huberman, A. M. (1994). Qualitative data analysis: An expanded sourcebook
(2nd ed.). SAGE.
Miles, M. B., Huberman, A. M., & Saldaña, J. (2020). Qualitative data analysis: A methods
sourcebook (4th ed.). SAGE.
Miller, D. C., & Salkind, N. J. (2002). Handbook of research design & social measurement
(6th ed.). SAGE.
Mollenhauer, K., & Uhlendorff, U. (1992). Sozialpädagogische Diagnosen: Über Jugendliche
in schwierigen Lebenslagen. Juventa.
Morgan, D. L. (2014). Integrating qualitative and quantitative methods: A pragmatic
approach. SAGE.
OECD (2007). OECD principles and guidelines for access to research data from public
funding. http://www.oecd.org/science/inno/38500813.pdf
O’Leary, Z. (2018). Research question. SAGE.
Oswald, H. (2010). Was heißt qualitativ forschen? Warnungen, Fehlerquellen,
Möglichkeiten. In B. Friebertshäuser, A. Langer, & A. Prengel (Eds.), Handbuch qualitative
Forschungsmethoden in der Erziehungswissenschaft (3rd ed., pp. 183–201). Juventa.
Peirce, C. S. (1974). Collected papers of Charles Sanders Peirce (C. Hartshorne &
P. Weiss, Eds.; 4th ed.). Belknap Press of Harvard University Press. https://
books.google.de/books?id=G7IzSoUFx1YC. (Original work published 1931.)
Popper, K. R. (2010). Die beiden Grundprobleme der Erkenntnistheorie: Aufgrund von
Manuskripten aus den Jahren 1930–1933 (T. E. Hansen, Ed.; 3rd ed.). Mohr Siebeck.
(Original work published 1979.)
Preisendörfer, P. (1999). Umwelteinstellungen und Umweltverhalten in Deutschland. VS
Verlag für Sozialwissenschaften. https://doi.org/10.1007/978-3-663-11676-9
Prosch, B., & Abraham, M. (2006). Gesellschaft, Sinn und Handeln: Webers Konzept des
sozialen Handelns und das Frame-Modell. In R. Greshoff & U. Schimank (Eds.),
Integrative Sozialtheorie? Esser – Luhmann – Weber (pp. 87–109). VS Verlag für
Sozialwissenschaften. https://doi.org/10.1007/978-3-531-90259-3_5
Przyborski, A., & Wohlrab-Sahr, M. (2014). Qualitative Sozialforschung: Ein Arbeitsbuch
(4th ed.). Oldenbourg Verlag.
Pürer, H. (2003). Publizistik- und Kommunikationswissenschaft: Ein Handbuch. UVK.
Rädiker, S., & Kuckartz, U. (2020). Focused analysis of qualitative interviews with MAXQDA:
Step by step. MAXQDA Press. https://doi.org/10.36192/978-3-948768072
Rehbein, J., Schmidt, T., Meyer, B., Watzke, F., & Herkenrath, A. (2004). Handbuch für das
computergestützte Transkribieren nach HIAT. https://www.exmaralda.org/hiat/files/azm_56.pdf
Ritchie, J., & Spencer, L. (1994). Qualitative data analysis for applied policy research. In
A. Bryman & R. G. Burgess (Eds.), Analyzing qualitative data (pp. 173–194). Taylor &
Francis. https://doi.org/10.4324/9780203413081_chapter_9
Ritchie, J., Spencer, L., & O’Connor, W. (2003). Carrying out qualitative analysis. In
Qualitative research practice: A guide for social science students and researchers
(pp. 219–261). SAGE.
Ritsert, J. (1972). Inhaltsanalyse und Ideologiekritik: Ein Versuch über kritische
Sozialforschung. Athenäum.
Rorty, R. (1979). Philosophy and the mirror of nature. Princeton University Press.
Rössler, P. (2017). Inhaltsanalyse (3rd ed.). UVK Verlagsgesellschaft mbH mit UVK/Lucius.
Rust, I. (2019). Theoriegenerierung als explizite Phase in der qualitativen Inhaltsanalyse:
Auf dem Weg zur Einlösung eines zentralen Versprechens der qualitativen
Sozialforschung. https://t1p.de/Theoriegenerierende-Inhaltsanalyse
Saldaña, J. (2013). The coding manual for qualitative researchers (2nd ed.). SAGE.
Schmidt, C. (2000). Analyse von Leitfadeninterviews. In U. Flick, E. von Kardorff, &
I. Steinke (Eds.), Qualitative Forschung: Ein Handbuch (pp. 447–455). Rowohlt
Taschenbuch Verlag.
Schmidt, C. (2010). Auswertungstechniken für Leitfadeninterviews. In B. Friebertshäuser
& A. Prengel (Eds.), Handbuch qualitativer Forschungsmethoden in der
Erziehungswissenschaft (3rd ed., pp. 473–486). Juventa.
Schreier, M. (2012). Qualitative content analysis in practice. SAGE.
Schreier, M. (2014). Varianten qualitativer Inhaltsanalyse. Ein Wegweiser im Dickicht der
Begrifflichkeiten. Forum Qualitative Sozialforschung / Forum: Qualitative Social Research,
15(1), Art. 18. https://doi.org/10.17169/fqs-15.1.2043
Schutz, A. (1962). Collected papers I. The problem of social reality. Martinus Nijhoff.
Seale, C. (1999a). The quality of qualitative research. SAGE.
Seale, C. (1999b). Quality in qualitative research. Qualitative Inquiry, 5(4), 465–478.
https://doi.org/10.1177/107780049900500402
Seale, C., & Silverman, D. (1997). Ensuring rigour in qualitative research. European
Journal of Public Health, 7(4), 379–384. https://doi.org/10.1093/eurpub/7.4.379
Silver, C., & Lewins, A. (2014). Using software in qualitative research: A step-by-step guide
(2nd ed.). SAGE.
Spencer, L., Ritchie, J., Lewis, J., & Dillon, L. (2003). Quality in qualitative evaluation: A
framework for assessing research evidence. https://www.gov.uk/government/
publications/government-social-research-framework-for-assessing-research-evidence
Sprenger, A. (1989). Teilnehmende Beobachtung in prekären Handlungssituationen: Das
Beispiel Intensivstation. In R. Aster, H. Merkens, & M. Repp (Eds.), Teilnehmende
Beobachtung: Werkstattberichte und methodologische Reflexionen (pp. 35–56). Campus.
Stamann, C., Janssen, M., & Schreier, M. (2016). Qualitative Inhaltsanalyse – Versuch einer
Begriffsbestimmung und Systematisierung. Forum Qualitative Sozialforschung / Forum:
Qualitative Social Research, 17(3), Art. 16. https://doi.org/10.17169/fqs-17.3.2581
Stamann, C., Janssen, M., Schreier, M., Whittal, A., & Dahl, T. (Eds.). (2020). FQS special
issue ‘Qualitative Content Analysis II’. Forum Qualitative Sozialforschung / Forum:
Qualitative Social Research, 21(1). http://www.qualitative-research.net/index.php/fqs/
issue/view/66