<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "https://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article article-type="research-article" xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="eng" dtd-version="1.1">
	<front>
		<journal-meta>
			<journal-id journal-id-type="eissn">2211-7253</journal-id>
			<journal-id journal-id-type="publisher-id">DuJAL</journal-id>
			<journal-title-group>
				<journal-title>Dutch Journal of Applied Linguistics</journal-title>
			</journal-title-group>
			<issn publication-format="print">2211-7245</issn>
			<issn publication-format="online">2211-7253</issn>
			<publisher>
				<publisher-name>Openjournals.nl</publisher-name>
			</publisher>
		</journal-meta>
		<article-meta>
			<article-id pub-id-type="doi">10.51751/DuJAL16248</article-id>
			<title-group>
				<article-title>Taming our wild data: On intercoder reliability in discourse research</article-title>
				<alt-title alt-title-type="running-head">Taming our Wild Data</alt-title>
			</title-group>
			<contrib-group>
				<contrib contrib-type="author" corresp="yes">
					<name name-style="western">
						<surname>van Enschot</surname>
						<given-names>Renske</given-names>
					</name>
					<email xlink:href="mailto:r.vanenschot@tilburguniversity.edu">r.vanenschot@tilburguniversity.edu</email>
					<xref ref-type="aff" rid="AFF000001"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Spooren</surname>
						<given-names>Wilbert</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000002"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>van den Bosch</surname>
						<given-names>Antal</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000003"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Burgers</surname>
						<given-names>Christian</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000004"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Degand</surname>
						<given-names>Liesbeth</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000005"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Evers-Vermeul</surname>
						<given-names>Jacqueline</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000003"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Kunneman</surname>
						<given-names>Florian</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000003"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Liebrecht</surname>
						<given-names>Christine</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000001"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Linders</surname>
						<given-names>Yvette</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000002"/>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Maes</surname>
						<given-names>Alfons</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000001"/>
				</contrib>
				<aff id="AFF000001">
					<label>1</label>
					<institution-wrap>
						<institution>Tilburg center for Cognition and Communication, Tilburg University</institution>
					</institution-wrap>
				</aff>
				<aff id="AFF000002">
					<label>2</label>
					<institution-wrap>
						<institution>Centre for Language Studies, Radboud University</institution>
					</institution-wrap>
				</aff>
				<aff id="AFF000003">
					<label>3</label>
					<institution-wrap>
						<institution>Institute for Language Sciences, Utrecht University</institution>
					</institution-wrap>
				</aff>
				<aff id="AFF000004">
					<label>4</label>
					<institution-wrap>
						<institution>Amsterdam School of Communication Research (ASCoR), University of Amsterdam</institution>
					</institution-wrap>
				</aff>
				<aff id="AFF000005">
					<label>5</label>
					<institution-wrap>
						<institution>Institute for Language and Communication, University of Louvain</institution>
					</institution-wrap>
				</aff>
			</contrib-group>
			<pub-date publication-format="online">
				<day>26</day>
				<month>03</month>
				<year>2024</year>
			</pub-date>
			<volume content-type="number">13</volume>
			<volume content-type="year">2024</volume>
			<fpage specific-use="PDF">1</fpage>
			<lpage>24</lpage>
			<history>
				<date date-type="received">
					<day>25</day>
					<month>07</month>
					<year>2023</year>
				</date>
				<date date-type="accepted">
					<day>02</day>
					<month>02</month>
					<year>2024</year>
				</date>
				<date date-type="pub">
					<day>26</day>
					<month>03</month>
					<year>2024</year>
				</date>
			</history>
			<permissions>
				<copyright-statement>Copyright 2024 by the author(s).</copyright-statement>
				<copyright-year>2024</copyright-year>
				<copyright-holder>the author(s)</copyright-holder>
				<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/" xlink:title="CC BY">
					<license-p>This is an open access article distributed under the terms of the CC BY 4.0 license.</license-p>
				</license>
			</permissions>
			<self-uri content-type="PDF" xlink:href="DuJAL16248_text.pdf"/>
			<abstract>
				<title>Abstract</title>
				<p>Many research questions in the field of applied linguistics are answered by manually analyzing data collections or corpora: collections of spoken, written and/or visual communicative messages. In this kind of quantitative content analysis, the coding of subjective language data often leads to disagreement among raters. In this paper, we discuss causes of and solutions to disagreement problems in the analysis of discourse. We discuss crucial factors determining the quality and outcome of corpus analyses, and focus on the sometimes tense relation between reliability and validity. We evaluate formal assessments of intercoder reliability. We suggest a number of ways to improve the intercoder reliability, such as the precise specification of the variables and their coding categories and carving up the coding process into smaller substeps. The paper ends with a reflection on challenges for future work in discourse analysis, with special attention to big data and multimodal discourse.</p>
			</abstract>
			<kwd-group kwd-group-type="uncontrolled">
				<title>Keywords</title>
				<kwd>intercoder reliability</kwd>
				<kwd>discourse</kwd>
				<kwd>quantitative content analysis</kwd>
				<kwd>complex discourse data</kwd>
				<kwd>hands-on procedures</kwd>
			</kwd-group>
			<funding-group>
				<funding-statement>No funding was received for this study.</funding-statement>
			</funding-group>
			<custom-meta-group>
				<custom-meta>
					<meta-name>Author contributions</meta-name>
					<meta-value>Renske van Enschot &amp; Wilbert Spooren: Conceptualization, Writing&#x00A0;&#x2013; original draft, Writing&#x00A0;&#x2013; review and editing, Supervision, Project administration. Antal van den Bosch, Christian Burgers, Liesbeth Degand, Jacqueline Evers-Vermeul, Florian Kunneman, Christine Liebrecht, Yvette Linders &amp; Alfons Maes: Conceptualization, Writing&#x00A0;&#x2013; original draft, Writing&#x00A0;&#x2013; review and editing.</meta-value>
				</custom-meta>
				<custom-meta>
					<meta-name>Statement of interest</meta-name>
					<meta-value>All authors declare that they do not have any competing interests.</meta-value>
				</custom-meta>
				<custom-meta>
					<meta-name>Statement of technology use</meta-name>
					<meta-value>No AI-based generative technology was used in the preparation of this manuscript and the execution of the research that the manuscript reports upon.</meta-value>
				</custom-meta>
			</custom-meta-group>
		</article-meta>
	</front>
	<body>
		<sec sec-type="heading-1">
			<label>1</label>
			<title>Taming our wild data: On intercoder reliability in discourse research</title>
			<p>Many research questions in the field of communication, linguistics and cognition are answered on the basis of manual analyses of data collections or corpora: collections of (transcribed) spoken, written and/or visual communicative messages. Although many different forms of corpus analysis are used (Krippendorff, 2019), the generic base can be defined as assigning interpretative categories to particular variables in the corpus. For example, particular words or expressions can be classified as having an intensifying meaning or as being ironic or metaphoric; gestures or pictures can be classified as representational or decorative; pitch patterns can be categorized as either expressing or lacking a feeling of knowing on the part of the producer; a particular interpretation of an artful poster or advertisement can be classified as &#x2018;matching the intended meaning&#x2019; or not; a coherence relation between two utterances in a discourse can be labeled as semantic or pragmatic.</p>
			<p>Obviously, these categorization tasks are extremely diverse in nature. For one thing, the span of the units can vary enormously: units can be sounds, words, phrases, utterances, images, complete texts or conversations. For another, the type of unit varies: annotating pitch in a telephone interaction is a completely different task from annotating the verbal language in that interaction, and the analysis of words is a task different from the analysis of gestures or images. It becomes even more complicated when the different modalities are combined. It is this variety in span and type of units that makes the analysis of discourse more intricate than the analysis that restricts itself to one linguistic level such as morphology and grammar. Despite the diversity, all of these activities fall under the heading of discourse coding: analyzing discourse in a systematic way in order to gain insight into underlying patterns. Often, this task is performed in a quantitative way, establishing or comparing frequencies of use, or trying to extrapolate frequencies from samples to populations.</p>
			<p>In this kind of quantitative content analysis of discourse, the coding of subjective language data often leads to disagreement among raters. This is partly due to coding errors, and partly due to the inherent ambiguity of the language phenomena. Disagreement can occur even after an extensive training phase, even if an explicit code book is used that has been tested and adapted, even if the number of coding categories is limited, and even if experts are employed instead of naive and untrained coders (Spooren &amp; Degand, 2010). Problems increase for the analysis of static and dynamic visuals in discourse. Often, several rounds of coding are necessary to reach a sufficiently high intercoder reliability statistic such as Cohen&#x2019;s Kappa. At the same time, our publication outlets require that such a Kappa is reached after only one round of coding (cf. Neuendorf, 2002, p.&#x00A0;146). Allegedly, only with one round of coding, categorizations can be considered replicable and valid.</p>
			<p>In most cases, the discourse units of interest come from a continuous stream of information. As a consequence, these units are often not predefined, and coders need to unitize the discourse before they can actually start categorizing these units. This may give rise to agreement issues as well, since coders can disagree about the number, length and alignment of the units in a discourse. For an analysis of the intricacies of unitizing issues and a valuable suggestion for quantitatively assessing the agreement on the unitizing effort, we refer the reader to Krippendorff (2019), Krippendorff et al. (2016) and Mathet et al. (2015). In the remainder of this paper, we focus on the issue of categorizing predefined units. Our restriction does not imply that we discard unitizing discourse, or the possibility that unitizing and categorization are interdependent, but we do believe that categorizing discourse units is a challenging task in itself, worth a separate discussion. Also worthy of a separate discussion is crowdsourcing. This can be seen as an interesting development that researchers make use of to have their data annotated by large numbers of naive coders, for example via ProLific (Suviranta &amp; Hiippala, 2022). Another noteworthy omission is the possibility of biases in cross-cultural analyses (Pe&#x00F1;a, 2007; Quiros-Ramirez &amp; Onisawa, 2015).</p>
			<table-wrap>
				<caption>
					<title>Table&#x00A0;1</title>
					<p>An overview of the key terms</p>
				</caption>
				<table frame="void" cellpadding="5">
					<thead>
						<tr>
							<th>
								<p>Term</p>
							</th>
							<th>
								<p>Definition</p>
							</th>
						</tr>
					</thead>
					<tbody>
						<tr>
							<td>
								<p>Coding categories</p>
							</td>
							<td>
								<p>The different values that a variable can take</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Corpus</p>
							</td>
							<td>
								<p>A collection of instances of discourse, usually enriched with metadata (for example about participants, register, source) and annotation (for example Part-of-Speech tagging, lemmatization)</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Discourse</p>
							</td>
							<td>
								<p>A continuous stream of naturally occurring communication in one or more modalities (for example written, spoken, video, computer-mediated) and using one or more verbal or non-verbal codes.</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Discourse coding</p>
							</td>
							<td>
								<p>The assignment of coding categories to a discourse unit</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Intercoder reliability (ICR)</p>
							</td>
							<td>
								<p>The degree to which different coders agree on assigning categories to discourse units</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Levels of measurement</p>
							</td>
							<td>
								<p>The nature of the information incorporated in a variable; for instance, an ordinal level of measurement allows for a rank order by which data can be sorted, but it does not allow for relative degree of difference between them</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Quantitative corpus analysis</p>
							</td>
							<td>
								<p>A type of research that uses corpus data and aims at finding frequency information about patterns in these corpora in order to be able to establish or compare those frequencies or extrapolate them to populations</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Unit</p>
							</td>
							<td>
								<p>The object of analysis to which a coder can assign a coding category</p>
							</td>
						</tr>
						<tr>
							<td>
								<p>Variable</p>
							</td>
							<td>
								<p>A property of interest that follows from the research question. It consists of at least two categories that presuppose a particular level of measurement</p>
							</td>
						</tr>
					</tbody>
				</table>
			</table-wrap>
			<p>In our paper, we make use of the terminology in Table&#x00A0;1. For example, in her <italic>quantitative corpus analysis</italic> of metaphoric language use in news discourse, Pasma (2011) worked with a <italic>corpus</italic> of newspaper texts from two time periods (1950/1951, 2002; 80,000 words per period). The <italic>units</italic> of her analysis were the words in the newspaper texts. For each word, she assessed the <italic>coding category</italic> of the <italic>variable</italic> metaphoricity: a word was used metaphorically or not; in other words, the <italic>level of measurement</italic> was nominal. She assessed the <italic>intercoder reliability (ICR)</italic> by having a second coder perform the same <italic>discourse coding</italic> task and calculating an ICR statistic.</p>
			<p>Our paper differs in several respects from related work like Bayerl and Paul (2011), who give a quantitative meta-analysis of factors influencing percentage agreement in a large number of linguistic corpus studies in different domains (word-sense disambiguation, prosodic transcriptions, and phonetic transcriptions). Our aim is to give a qualitative analysis of the factors involved, using case studies from various areas of discourse analysis. As a consequence, we are not limited&#x00A0;&#x2013; as Bayerl and Paul are&#x00A0;&#x2013; to the restricted number of factors available in the data set that is used.</p>
			<p>In the next sections, we first sketch the scope of the reliability problem by describing different degrees of &#x2018;wildness&#x2019; in discourse data. We then describe the tension between reliability and validity. Subsequently, we provide an overview of shortcomings of using Cohen&#x2019;s Kappa scores as a measure of intercoder reliability&#x00A0;&#x2013; ICR from now on&#x00A0;&#x2013;, and suggest alternative statistical ICR metrics. Our paper continues with hands-on advice on how to improve ICR, and concludes with a reflection on where to go from here, with special attention to big data and multimodality.</p>
		</sec>
		<sec sec-type="heading-1">
			<label>2</label>
			<title>Different degrees of &#x2018;wildness&#x2019; in discourse data</title>
			<p>The quality and outcome of corpus analyses, as well as the success of ICR tests, is determined by a large number of factors. In this section, we will discuss the effects of three important factors on the openness and ambiguity (&#x2018;wildness&#x2019;) of discourse data: different ways of collecting data, different ways of establishing coding categories, and the kind of coding categories that are taken into account.</p>
			<p>First, data collections can be elicited experimentally or selected from naturally occurring data. The latter tends to be more difficult to code reliably. For instance, Arts et al. (2011) asked their participants to produce referring expressions describing entities on a screen in a controlled setting. The result was an easy-to-code dataset of referential expressions containing only attributes visible on the screen. This differs sharply from the problems encountered while encoding the stylistic elements in naturally-occurring newspaper and web texts, such as the ones reported in Liebrecht (2015).</p>
			<p>Second, coding categories can be established in different ways. On the one hand, they can have a predefined theoretical position and definition. Examples are using an enchiridion&#x00A0;&#x2013; a short handbook&#x00A0;&#x2013; to categorize intensifiers on a lexical base (Van Mulken &amp; Schellens, 2012), or using an unambiguous and theory-based definition of verbal irony (Burgers et al., 2011). Low intercoder reliability could be prevented if researchers would study only variables with such well-defined categories that hardly cause any interpretation noise. On the other hand, coding categories can start from an intuition and emerge gradually and exploratively as the analysis proceeds (Van Enschot &amp; Donn&#x00E9;, 2013). Many interesting questions in the field of human communication are not ready for controlled types of research, and therefore require an explorative approach. Examples of such intriguing questions are: &#x2018;What makes a visual message metaphorical (Forceville &amp; Urios-Aparisi, 2009)?&#x2019; and &#x2018;Which linguistic or audio-visual cues can we consider to be deceptive (Hancock et al., 2007)?&#x2019; The coding of the latter type of data tends to be much more difficult than the coding of the first type.</p>
			<p>Note that agreement tests can be useful both in controlled and exploratory studies. Controlled studies require a high degree of precision and validity, and consequently only high levels of ICR outcomes are acceptable. In explorative studies, however, ICR scores can be used as a heuristic tool to objectify or specify individual intuitions, to try out coding category definitions or segmentation options in data collections. During this incremental process, lower ICR rates are acceptable, although one may want to eventually perform a more controlled analysis to validate the final version of the newly developed coding system.</p>
			<p>Third, a set of coding categories can vary from closed to open. On the one extreme, the coding categories form a dyadic, mutually exclusive, closed set (for example yes-no, present-absent), clearly defined in terms of objective characteristics. On the other extreme, the number of coding categories is not fixed (for example ways to intensify an utterance) or the coding categories are not mutually exclusive (for example literal versus figurative language). Such coding categories leave room for an exploratory analysis, but at the same time pose problems for ICR.</p>
			<p>Our discussion of these three factors illustrates that analytical data can be less or more wild, and that the degree to which our data are wild depends on the choices made by the researchers. The researchers make decisions about the type of questions asked (for example, do they focus on the question whether or not an utterance is subjective or whether a specific word indicating subjectiveness such as <italic>terrific</italic> occurs in the utterance), about the collection method used (for example, do they focus on a corpus of all tweets produced in a one-hour slot in The Netherlands, or on the one-word-production results of a word-listing task), about the theoretical framework and position taken (for example, do they start from the position that all subjective words can be divided into two or three major coding categories or not), and about the way in which theory is operationalized into coding categories (for example, do they define subjective words as a closed set of identifiable elements in discourse, or as an open-ended class).</p>
			<p>These choices to a large extent determine the success of the coding task. Once we agree that data can be wild, the question is which precautions can be taken to make intercoder reliability tasks doable, not trivial, and successful or at least informative. We will address such precautions in the section on how to improve intercoder reliability.</p>
		</sec>
		<sec sec-type="heading-1">
			<label>3</label>
			<title>Reliability versus validity</title>
			<p>Data quality is a matter of both reliability and validity. To illustrate the difference between reliability and validity, various scholars use the metaphor of a target or a dart board (Krippendorff, 2019; Trochim, 2006). Imagine your goal is to systematically hit the bullseye. Your performance is unreliable if the darts are randomly spread across the board. It is reliable if you consistently hit the same spot on the board&#x00A0;&#x2013; irrespective of whether this is the bullseye or some other spot, such as the &#x2018;double 20&#x2019;. However, your performance is only valid if you consistently hit the intended spot on the board, and not if you just hit the bullseye every now and then. Thus, the &#x2018;dart board&#x2019; metaphor demonstrates that reliability and validity both need to be high, and that high reliability does not necessarily imply high validity. Similarly, aiming for high ICR scores is not a guarantee for good validity. On the contrary, it may even yield serious problems for the construct, external and internal validity of the research.</p>
			<p>An established viewpoint is that reliability is at least a necessary condition for validity (Moss, 1994). What does this imply for researchers with &#x2018;wild&#x2019; data that frequently result in insufficient intercoder reliability scores? One way out of this problem is to consider the theory or the coding procedure yielding the analysis of such unreliable data to be underdeveloped to such a degree that the researchers should go back to the drawing board. Alternatively, we could restrict the generalizability of our results to the limited set of data that we can code reliably. Such a solution is chosen by Liebrecht (2015) for the analysis of intensified language. She reports analyses on the subset of data on which both coders agreed. Of course, this limits the generalizability of the results. To accommodate that problem, she also reports findings for the intensifiers identified by each coder separately, and only draws firm conclusions when all results point in the same direction.</p>
			<p>Below, we discuss the different types of validity, with special attention for how validity can be at odds with reliability. It will become clear that a high ICR&#x00A0;&#x2013; if achievable&#x00A0;&#x2013; is not always a guarantee for having valid data.</p>
			<sec sec-type="heading-2">
				<label>3.1</label>
				<title>Construct validity</title>
				<p>Construct validity refers to the question whether the coded data accurately reflect the theoretical constructs they are supposed to measure (Elmes et al., 2012, pp.&#x00A0;185-187). Aiming for high reliability scores may be at odds with this type of validity. For example, Wallace (2015) argues that the pragmatic context of utterances is a common feature of modern theoretical approaches to irony and sarcasm, but that this aspect is largely ignored in the ways in which various computational linguists have operationalized irony and sarcasm in their corpus-based research. In efforts that guarantee a boost in reliability, many researchers have coded sarcasm by looking for specific words or word combinations, such as variants of the word sarcasm (sarcastic, sarcastically, etc.), or words that are often used to mark sarcasm (for example <italic>Yeah right</italic>). Wallace (2015) therefore qualifies such automatic identification procedures based on word usage as &#x201C;shallow&#x201D;, because they do not take into account semantic and pragmatic information about the speaker or situation, and are likely to miss many instances of sarcasm.</p>
				<p>For instance, an utterance such as &#x2018;<italic>Gerald Ford was a great president. Yeah right</italic>&#x2019; is more likely to be interpreted as literal when said by a supporter of the Republican Party, and as sarcastic when it comes from a supporter of the Democrats. Wallace thus calls upon computational linguists to develop more advanced computational models that take into account not only syntactic aspects, but also semantic and pragmatic aspects. Thus, even if identification procedures achieve satisfactory or high reliability scores&#x00A0;&#x2013; coding instances of <italic>Yeah, right</italic> in a corpus can easily be done very reliably&#x00A0;&#x2013; it is important to assess critically whether the data analysis meets the criterion of construct validity, i.e., whether it captures all cases included in the theoretical definition of the variable.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>3.2</label>
				<title>External and internal validity</title>
				<p>External validity refers to the way in which observations can be generalized to other situations outside of the specific data investigated. The internal validity criterion invites the researcher to search for confounding factors. Both kinds of validity can be at odds with reliability. An example comes from research on the quality of the spelling in students&#x2019; writing. It makes an enormous difference whether the researcher analyzes the spelling errors in dictations, or in texts that are composed by students themselves. As Van den Bergh et al. (2011, p.&#x00A0;6) point out, analyzing spelling errors in the former text types can be done in a very reliable way (they report 100&#x202F;% intercoder reliability for dictations). However, there are issues of external validity. Although dictations can give a systematic picture of what children are capable of in terms of specific spelling difficulties, children&#x2019;s numbers of spelling errors in dictations are not predictive of the number of spelling errors in their own writing: Van den Bergh et al. report correlations between .11 and .17 (2011, p.&#x00A0;12). Next to external validity, the internal validity is at stake in this analysis as well. First, the number of errors in students&#x2019; own writing may be determined, at least in part, by their proficiency in coming up with an alternative formulation, allowing them to avoid words that are difficult to spell. Second, variation in numbers of spelling errors in dictations and students&#x2019; own texts may also be attributed to differences in task. If students take dictations, their main focus is on form, not on content. However, if students write their own texts, their focus is on content, and less so on correctness of form. This confounding of factors makes it hard to compare the outcomes of studies in which different tasks are used, and it illustrates that reliability is not a sufficient condition for validity.</p>
			</sec>
		</sec>
		<sec sec-type="heading-1">
			<label>4</label>
			<title>How to assess intercoder reliability statistically</title>
			<p>In this section, we turn to a selection of settings in which the most widespread intercoder reliability metric, Cohen&#x2019;s Kappa, can be misleading: when each item is annotated by a different selection of annotators, when the categories are in an ordering relation, or when there is an imbalance in the frequency at which different categories are annotated. We provide an overview of metrics that can be applied alternatively to Cohen&#x2019;s Kappa. We will focus on the main characteristics and advantages of these alternative metrics, without providing formulas.<xref ref-type="fn" rid="FN000001"><sup>1</sup></xref> The issues we describe all assume that the discourse units have already been established, that is: we describe the situation in which the unitizing stage has been finished, and the researcher has to categorize the units of interest.</p>
			<p>Cohen&#x2019;s Kappa was proposed by Jacob Cohen (1960) and takes into account the prior chance that two annotators agree on the annotation of any coding category. This makes Cohen&#x2019;s Kappa a more realistic metric for intercoder reliability than percentage agreement, which can easily give misleading insights. For instance, one study using two coding categories has a fifty percent chance that coders agree, while another study using four coding categories yields a 25 percent agreement chance. As a result, given a similar percentage agreement, the first study will yield higher ICR scores than the second study simply due to chance (Artstein &amp; Poesio, 2008, pp.&#x00A0;558-559).</p>
			<p>As a way to account for chance agreement, Cohen&#x2019;s Kappa presumes all items in a set to be coded by the same two annotators who are considered statistically independent. The chance for annotator 1 to choose certain categories is calculated, as is the chance for annotator 2 to choose certain categories. Based on these chances, the possibility for two annotators to agree coincidentally is calculated. As Krippendorff (2011) explains, this is not a realistic perspective on the role of chance in ICR: Annotators are dealing with the same categories to code and have read the same code book, and hence their annotations should not be seen as independent sets. A consequence of this is that the Cohen&#x2019;s Kappa metric does not penalize patterns where the class distributions of the annotators are highly distinct. For example, if annotator 1 has chosen category A for 80&#x202F;% of the items and annotator 2 has only chosen this category for 40&#x202F;% of the items, the chance that both agree on this category is (0.8*0.4=) 0.36, and the chance that the two agree on the other category is (0.2*0.6=) 0.12, resulting in a chance agreement of (0.36&#x202F;+&#x202F;0.12=) 0.48. This is a smaller penalty than when both annotators would for example choose category A for 40&#x202F;% of the times (a change agreement of 0.52). It would make more sense to look at the combined decisions of all annotators as a proxy of the class distribution, which is done in metrics like Fleiss&#x2019; Kappa (Fleiss, 1971), Scott&#x2019;s pi (Scott, 1955) and Krippendorff&#x2019;s Alpha (Hayes &amp; Krippendorff, 2007; Krippendorff, 2019). These metrics consider the proportion of times that each coding category is chosen by annotators, as well as the agreement per single item. Essentially, annotators are then seen as interchangeable, which is more in line with the notion of reliability of annotations: if the annotations in the end rely on a straightforward decision procedure which is reflected in the agreement, it should not matter who does the annotations and any one annotator could be replaced by another without much effect on the agreement.</p>
			<p>An important property of the coding task is the measurement level of the variable. In standard form, the Cohen&#x2019;s Kappa metric presumes a nominal scale, in which no ordering exists between the coding categories. It will give wrong insights when applied to other than nominal variables, with ordinal, interval or ratio levels. With these levels of measurement, it should be penalized when two coders annotate coding categories that have a larger distance to one another. Metrics that do apply such a penalization are the Weighted Cohen&#x2019;s Kappa (Cohen, 1968) and Krippendorff&#x2019;s Alpha (Hayes &amp; Krippendorff, 2007; Krippendorff, 2019). In these metrics, the agreement, or disagreement in the case of Krippendorff&#x2019;s Alpha, for any pair of levels is weighted by taking into account the distance between the categories, such that more distanced categories add a lot less to the agreement score, or a lot more to the disagreement score. Krippendorff&#x2019;s Alpha also allows for missing data points, by considering the total number of annotations that were made.</p>
			<p>Another factor that needs to be considered when assessing intercoder reliability is data skew: the degree to which data is asymmetrically distributed over coding categories. Di Eugenio and Glass (2004) and Jeni et al. (2013) show that Cohen&#x2019;s Kappa and Krippendorff&#x2019;s Alpha are highly sensitive to imbalanced variables: the agreement score will drastically decrease with a bigger data skew. The reason is that the prior chance of annotators making similar annotations is high if there are one or a few dominant coding categories. As the higher chance agreement is subtracted from the percentage agreement, a high percent agreement may still result in a relatively low ICR. In other words, it is unlikely to reach a high Cohen&#x2019;s Kappa or Krippendorff&#x2019;s Alpha when most of the items fall under one coding category. A solution is to calculate the Kappa Max (Umesh et al., 1989), which returns a Kappa value that is relative to the upper bound of the Kappa that follows the strict chance agreement. Alternatively, adaptations have been suggested to correct Kappa for such biases, such as PABAK (prevalence-adjusted bias-adjusted kappa; Byrt et al., 1993).</p>
			<p>The Mutual F-score is another metric that can be used to mitigate the influence of category imbalance. Mutual F-scores are based on F-scores, which are often used in Information Retrieval and Machine Learning for system evaluation (Van Rijsbergen, 1979). In contrast to the other metrics, the Mutual F-score estimates the agreement per coding category rather than over the total of annotations. Considering the annotations of one coder as ground truth, for each category the annotations of a second coder can be assessed with respect to this ground truth. At the basis of the F-score are three counts: of True Positives (TP), False Positives (FP) and False Negatives (FN). TP&#x202F;s are instances that both coders annotate with the coding category in focus, FP&#x202F;s are instances that are only annotated by the second coder with the coding category in focus and FN&#x202F;s are instances that are only annotated by the first coder with the coding category in focus. Based on these values, Precision can be calculated as the percentage of instances coded by the second annotator with the coding category in focus, that were actually correct with respect to the ground truth annotations of the first coder (TP&#x202F;s divided by the total of TP&#x202F;s and FP&#x202F;s). Recall can be calculated as all instances that were annotated by the first annotator with the coding category in focus, that were annotated in the same way by the second annotator (TP&#x202F;s divided by the total of TP&#x202F;s and FN&#x202F;s). The F1 score is the harmonic mean of Precision and Recall: 2*(Precision*Recall)/(Precision+Recall). To reach the Mutual F-score between annotators, the role of the ground truth annotator is swapped, and the F-score is computed again in the other direction. Then, Mutual F-score can be computed as the mean of the two directional F-scores. As the annotations of other categories by both coders are left out of this calculation (the TN&#x202F;s, true negatives), the Mutual F-score allows to zoom in on the agreement per coding category.</p>
			<p>To illustrate the informativeness of the Mutual F-score in comparison to other metrics, we provide an example from our own work. Kunneman et al. (2015) aimed to automatically detect sarcasm in Twitter posts, and trained a model based on tweets that are accompanied by Dutch hashtags likely deployed by the sender to signify the sarcasm in their post. To validate this approach, for a random sample of 250 messages that include such a hashtag, three annotators coded whether they actually were sarcastic (Kunneman et al., 2015, p.&#x00A0;505). The confusion matrix for one of the coder pairs is displayed in Table&#x00A0;2.</p>
			<table-wrap>
				<caption>
					<title>Table&#x00A0;2</title>
					<p>Example confusion matrix on coding Twitter posts as sarcastic or non-sarcastic, as part of the study by Kunneman et al. (2015)</p>
				</caption>
				<table frame="void" cellpadding="5">
					<thead>
						<tr>
							<th>
								<p></p>
							</th>
							<th>
								<p>Coder B: Sarcastic</p>
							</th>
							<th>
								<p>Coder B: Non-sarcastic</p>
							</th>
						</tr>
					</thead>
					<tbody>
						<tr>
							<td>
								<p><bold>Coder A: Sarcastic</bold></p>
							</td>
							<td>
								<p>205</p>
							</td>
							<td>
								<p>12</p>
							</td>
						</tr>
						<tr>
							<td>
								<p><bold>Coder A: Non-sarcastic</bold></p>
							</td>
							<td>
								<p>11</p>
							</td>
							<td>
								<p>22</p>
							</td>
						</tr>
					</tbody>
				</table>
			</table-wrap>
			<p>The percentage agreement between the two coders is 91&#x202F;% ((205&#x202F;+&#x202F;22)/250), but due to a dominance of the coding category <italic>Sarcastic</italic>, both Cohen&#x2019;s Kappa and Krippendorff&#x2019;s Alpha are 0.60. Considering the annotations of coder A as ground truth, the F-score for the coding category <italic>Sarcastic</italic> is 0.95 (TP = 205, FP = 11, FN = 12, Precision = 205/(205&#x202F;+&#x202F;11) = 0.95, Recall = 205/(205&#x202F;+&#x202F;12) = 0.95, F1 = 2*((0.95*0.95)/(0.95&#x202F;+&#x202F;0.95)) = 0.95) and the F-score for the coding category <italic>Non-sarcastic</italic> is 0.66 (TP = 22, FP = 12, FN = 11, Precision = 22/(12&#x202F;+&#x202F;22) = 0.65, Recall = 22/(11&#x202F;+&#x202F;22) = 0.67, F1 = 2*((0.65*0.67)/(0.65&#x202F;+&#x202F;0.67)) = 0.66). Given that the dataset of this variable contains two coding categories, the Mutual F-score on the coding category <italic>Sarcastic</italic> is therefore 0.95, and for <italic>Non-sarcastic</italic> it is 0.66. These two values highlight the difference in agreement on the two categories, where the less frequent coding category is the more problematic.</p>
			<p>After having arrived at a reliability metric, the next step is interpreting this metric. Earlier on, we suggested that the ICR depends on the nature of the data and the research objective(s). This suggestion would imply that an ICR metric should be interpreted relative to these objectives. Interpretation may vary per type of study: for hypothesis testing, a higher ICR is required than for explorative studies. Similar suggestions can be found in Grove et al. (1981) and Spooren and Degand (2010) (see also the discussion in Artstein &amp; Poesio, 2008, and Mathet et al., 2012).</p>
			<p>In light of these examples, it is important to fully understand the mechanisms of a metric when interpreting its outcomes and be aware of how Cohen&#x2019;s Kappa can mislead. Next to the overview in this section, Artstein and Poesio (2008) give an excellent, elaborate overview of the mathematics and assumptions behind the ICR metrics, and discuss issues of annotator bias (not all coders use each category equally often) and prevalence bias (some categories are used more often than others). Such overviews enable researchers to interpret the ICR metrics correctly, including Cohen&#x2019;s kappa, allowing them &#x201C;to make an informed choice regarding the coefficients they use for measuring agreement&#x201D; (Artstein &amp; Poesio, 2008, p.&#x00A0;590). The discussion also shows that existing heuristics for interpreting outcomes that do not take into account the context of annotation, seem too simplistic. In line with Perreault and Leigh (1989) who state that &#x201C;different indices reflect different aspects of reliability&#x201D; (p.&#x00A0;146), we recommend assessing interrater agreement with several metrics. This way, we can achieve a more complete interpretation of the factors in play.</p>
		</sec>
		<sec sec-type="heading-1">
			<label>5</label>
			<title>How to improve intercoder reliability</title>
			<p>Researchers need hands-on advice and practical solutions to ICR problems. In this section, we meet this need and suggest a number of concrete ways to improve intercoder reliability: making categories independent (5.1), reducing the number of categories (5.2), decomposing the analytical process into smaller steps (5.3) and some procedural measures such as optimizing the coding instructions (5.4). Some of these topics are also discussed by Bayerl and Paul (2011), who used a quantitative meta-analysis to identify several factors influencing reported agreement values (like number of annotators and number of categories in a coding scheme). Our discussion is more qualitative, allowing us to discuss factors not touched upon by Bayerl and Paul such as the decomposition of the coding process and the independence of the categories. Bayerl and Paul&#x2019;s discussion is limited to three relatively local domains (word-sense disambiguation, prosodic transcription, phonetic transcription), whereas our discussion focuses on the level of discourse data. Other useful sources are for instance Krippendorff (2019), Neuendorf (2002) and Selvi (2020).</p>
			<sec sec-type="heading-2">
				<label>5.1</label>
				<title>Make your categories independent</title>
				<p>Agreement success is highly determined by the relation between the categories within a given variable. Two major conditions are relevant here: one is whether the categories are mutually exclusive or not, the other is whether the categories are hierarchically ordered. The same unit of analysis can have different functions or interpretations, which may result in scalar categories or in units belonging to more than one of the categories within a given variable (for example, clauses can have different types of relations with other clauses). Obviously, reaching satisfactory agreement scores is easier if categories are mutually exclusive.</p>
				<p>Scalar variables can be transformed into binary variables to make the categories mutually exclusive. For example, in a study of the subjectivity of adjectives preceding or following causal connectives, Hoek et al. (2021) used the subjectivity scores on a continuous scale from 0 to 1 that were available in the so-called gold1000 lexicon of subjective adjectives (De Smedt &amp; Daelemans, 2012). For the sake of the analysis, explicit boundaries were used to create subsets of objective adjectives (defined as adjectives with a subjectivity score &lt; .20) and subjective adjectives (adjectives with a subjectivity score &gt; .70). The other adjectives were considered ambiguous and therefore excluded from the analysis.</p>
				<p>Categories can also be hierarchically ordered: coders are first asked to determine major categories and then asked to subclassify units within the assigned coding category. An example is coding discourse relations first in terms of semantic versus pragmatic relations, and then the exact relation type within the assigned coding category. Agreement success is endangered if coders have to apply such embedded coding tasks (cf. Scholman et al., 2016).</p>
				<p>This issue occurred in a study by Zufferey and Degand (2013) who report percentage agreements of three types of multilingual discourse relation annotation differing in the amount of specificity. For the most generic type of annotation, i.e., distinguishing between four categories of discourse relations (temporal, comparison, contingency, expansion, cf. Webber et al., 2019), agreement is highest, above 90&#x202F;%. For the second type, which subdivides for example contingency into condition and cause, agreement drops to 60-72&#x202F;%. The third, most specific type splits for example causal into reason and result, and yields agreement percentages between 39&#x202F;% and 53&#x202F;%. Part of the disagreement concerning the second and third type is caused by disagreements concerning the first, more generic type. This is because decisions regarding this first type directly impact decisions that have to be made for the second and third type; the latter decisions are dependent on the decisions for the first type. In example (1), the relation conveyed by <italic>when</italic> could arguably be either temporal or contingent. Disagreement regarding this first, more generic type automatically induces disagreement regarding the more specific type because the available subcategories will be different.</p>
				<disp-quote>
					<p>(1) The clich&#x00E9; of a Mediterranean lolling in the sun has become a mental reflex when trying to explain the cause of the crisis in the Eurozone.</p>
				</disp-quote>
				<p>How would one estimate the intercoder reliability in such a case? Suppose that intercoder reliability scores are calculated for the most specific type (such as whether the relation is hypothetical or general). To obtain these scores, only the cases are used on which the coders already agreed when looking at the more generic type (i.e., they agreed that the relation was one of condition and not of cause; they also agreed that it was a contingency relation, and not a temporal relation). A high ICR score for a more specific type implies a high ICR score for its overarching type. By contrast, if an ICR score for a more specific type is low, it remains to be established where the disagreement comes from. A recommendation is to proceed stepwise, to start analyzing at the first, most generic type, determine the ICR, and then move on to the second and then the third type, analyzing just the instances of the more generic type that were agreed upon. This makes it possible to situate the coding problems with more precision.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>5.2</label>
				<title>Reduce the number of categories</title>
				<p>Reducing the number of categories in a coding system is likely to improve intercoder reliability (see Bayerl &amp; Paul, 2011), but this is usually undesirable because a reduced coding system yields less information. An obvious first check is whether all categories are really needed. For example, Van Enschot and Hoeken (2015) originally had two subcategories in their analysis of tropes&#x00A0;&#x2013; a type of rhetorical figures&#x00A0;&#x2013; in TV commercials: one in which the verbal part of the TV commercial explicitly mentioned the trope, as in <italic>This woman is as beautiful as a rose</italic>, and one in which the verbal part did not address this link explicitly, as in <italic>This woman is beautiful</italic>. Both were regarded as explicit explanations of the trope, and during the final phase of the analysis were therefore combined into a single coding category, resulting in higher ICR scores.</p>
				<p>An advantage of reducing the number of categories per variable is that the remaining categories occur more frequently, which often avoids the statistical bias of unequal distribution. A case in point is the coding of the syntactic class of discourse markers (Bolly et al., 2014). This class of linguistic expressions is very heterogeneous, consisting mostly of coordinate and subordinate conjunctions such as <italic>but</italic> and <italic>because</italic>, and of adverbials such as <italic>well</italic> and <italic>actually</italic>, but also of less frequent members such as parentheticals (<italic>I mean</italic>, <italic>I think</italic>) or adjectives (<italic>first</italic>, <italic>good</italic>). This results in a variable with many categories, of which some occur infrequently. Depending on the general theory and the research question at hand, researchers can question whether it is useful to keep all possible syntactic categories or whether some of them should be grouped. Should they maintain fine-grained distinctions such as the one between coordinate and subordinate conjunctions, or between prepositions and prepositional phrases, or should they, for instance, choose to distinguish only the most probable syntactic classes (for example adverbials, conjunctions and prepositional phrases) and group all other possibilities in one encompassing other coding category, or even retain only two coding choices (for example between conjunctive and non-conjunctive). Some of these options are illustrated in Table&#x00A0;3.</p>
				<p>Let us assume that the coders have a data set of 50 occurrences to annotate. If they choose to code according to option 1 in Table&#x00A0;3 (with ten categories), an equal distribution of all categories would lead to a maximum of five occurrences per coding category. Now, knowing that adjectives or nouns used as discourse markers are very rare in English, it is highly probable that these categories will receive zero counts. This may lead to biases in the statistical analysis as demonstrated by Feinstein and Cicchetti (1990). Therefore, either the sample has to be increased to account for rare events, or the number of categories has to be reduced. Still another recommendation is to use a statistical measure such as Kappa Max or PABAK (Prevalence-Adjusted Bias-Adjusted Kappa) that is sensitive to uneven distributions (see the section on how to assess ICR statistically).</p>
				<table-wrap>
					<caption>
						<title>Table&#x00A0;3</title>
						<p>Coding options for the variable &#x2018;syntactic class&#x2019; of discourse markers</p>
					</caption>
					<table frame="void" cellpadding="5">
						<thead>
							<tr>
								<th>
									<p>Syntactic class 1</p>
								</th>
								<th>
									<p>Syntactic class 2</p>
								</th>
								<th>
									<p>Syntactic class 3</p>
								</th>
							</tr>
						</thead>
						<tbody>
							<tr>
								<td>
									<p>clause</p>
								</td>
								<td>
									<p>adverbial</p>
								</td>
								<td>
									<p>conjunctive</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>verbal phrase</p>
								</td>
								<td>
									<p>conjunction</p>
								</td>
								<td>
									<p>non-conjunctive</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>adverb</p>
								</td>
								<td>
									<p>prepositional</p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>coordinating conjunction</p>
								</td>
								<td>
									<p>other</p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>subordinating conjunction</p>
								</td>
								<td>
									<p></p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>adjective</p>
								</td>
								<td>
									<p></p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>preposition</p>
								</td>
								<td>
									<p></p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>prepositional phrase</p>
								</td>
								<td>
									<p></p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>noun</p>
								</td>
								<td>
									<p></p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>interjection</p>
								</td>
								<td>
									<p></p>
								</td>
								<td>
									<p></p>
								</td>
							</tr>
						</tbody>
					</table>
				</table-wrap>
			</sec>
			<sec sec-type="heading-2">
				<label>5.3</label>
				<title>Decompose the analytical process into smaller steps</title>
				<p>If reducing the number of categories is not possible without losing too much information, an interesting alternative is to decompose the analytical process into smaller, simpler steps (see Figure&#x00A0;1; see also Fort et al., 2012). Thus, instead of reducing the number of categories to be coded, one can simplify the coding decisions by decomposing coding steps, creating a hierarchical coding system, while at the same time reducing the number of categories that need to be considered during each step. The net result is that the same number of coding categories will be considered. The main advantage of this procedure is that the decision process is split up into smaller decision trees. Note that the decision tree described in Figure&#x00A0;1b seems to suggest a hierarchical variable and not the nominal variable described in Figure&#x00A0;1a. The same bias might hold for a categorization task as described in Figure&#x00A0;1a if the coder always starts with considering category A. To avoid the bias that category A is always considered first, the categories should, if possible, be presented to the coder in a random order.</p>
				<p>For example, in a research project on literary criticism (Linders, 2014), coders had to indicate which characteristics of novels were being evaluated by critics. For each evaluative statement, the coders had to choose which of the fifteen listed types of characteristics applied. Linders&#x2019; task consisted of making coders choose one of these fifteen options at once. An alternative would have been that the coders were confronted with a series of fourteen binary decisions: the first step might have been deciding whether the statement was about the book itself, the second about the effect the book had on the reader, the third about the book in relation to the world, et cetera. The result would have been a coding procedure that has a better chance of achieving a high ICR because each coding decision would be a relatively simple binary one. All in all, decomposing looks like a promising strategy, although it may be more time-consuming than a more straightforward single-step coding procedure.</p>
				<fig-group orientation="portrait" position="anchor">
					<fig orientation="portrait" position="anchor">
						<caption>
							<title>Figure&#x00A0;1a</title>
							<p>Schematic analytical process of a complex coding category</p>
						</caption>
						<graphic xlink:href="DuJAL16248_i0001.png"/>
					</fig>
					<fig orientation="portrait" position="anchor">
						<caption>
							<title>Figure&#x00A0;1b</title>
							<p>Simplified version of the analysis in Figure&#x00A0;1a</p>
						</caption>
						<graphic xlink:href="DuJAL16248_i0002.png"/>
					</fig>
				</fig-group>
			</sec>
			<sec sec-type="heading-2">
				<label>5.4</label>
				<title>Procedural measures</title>
				<p>In order to facilitate the coding procedure, a number of elements have to be taken into account. We will discuss the number of coders involved, the instructions given to the coders, and their training.</p>
				<sec sec-type="heading-3">
					<label>5.4.1</label>
					<title>Consider the number of coders</title>
					<p>An increase in the number of coders has been shown to correlate negatively with the agreement (Bayerl &amp; Paul, 2011). Two or three coders are standard in most studies in the field of communication, linguistics and cognition (for example Kunneman et al., 2015; Van Enschot &amp; Donn&#x00E9;, 2013; Van Mulken &amp; Schellens, 2012). If coding is relatively simple (for example with a few categories to be assigned in well-defined units), one additional coder who recodes part of the data is considered sufficient (for example Mol et al., 2012). If data are wilder or categories more diffuse, coding by three or more coders may be useful, not only to obtain a reliable analysis, but also to gradually develop a sufficient understanding of the research phenomenon. In general, including more coders is more reliable (Potter &amp; Levine-Donnerstein, 1999), but practical considerations make two or three coders reasonable.</p>
					<p>Another interesting strategy is to deploy several coders on a part of the data, to assess the intercoder reliability on this part with an agreement measure, and if the value is high, to deploy only one coder (not necessarily the same one) for the rest of the data. An example can be found in Martin et al. (2014), where one third of the data was coded by two coders, and the remaining two thirds were spread over both of them once the intercoder reliability had been assessed high enough.<xref ref-type="fn" rid="FN000002"><sup>2</sup></xref></p>
				</sec>
				<sec sec-type="heading-3">
					<label>5.4.2</label>
					<title>Optimize the coding instruction</title>
					<p>The coding instruction also plays an important role. How specific is the instruction? In the majority of cases, written instructions are used. Such instructions differ in specificity: some only present a description of the task; others contain various examples of the phenomenon under investigation with the risk that coders are biased by those examples. Frequently, coders get the opportunity to ask the researcher for further explanation after they have read the instruction. Then, the coders analyze the materials with the instruction in mind. A risk of this procedure is that coders gradually start leaving out certain analytical steps because of tiredness or subconscious automatic behavior. Possibly, a clearer and more ordered way of instructing the coders and guiding them through the analytical process is to not only let them read a written instruction, but also present the analytical procedure step by step in a decisional flowchart, like Burgers et al. (2011, 2016) and Scholman et al. (2016) did. With the decisional flowchart at hand, coders can follow the analytical process step by step, which prevents tiredness and automaticity.</p>
				</sec>
				<sec sec-type="heading-3">
					<label>5.4.3</label>
					<title>Train the coders</title>
					<p>The final factor involving the coding procedure is the degree to which coders are trained. Bayerl and Paul (2011, pp.&#x00A0;713-714) report a strong positive correlation between training and agreement scores, particularly if the intensity of training is high.</p>
					<p>Coders can have different degrees of expertise, based on the amount of training and experience. They can be not trained at all, having read the instruction by themselves, they can practice the instruction with a single text, or they can be trained with multiple texts and feedback rounds with the researcher. Coders can also be experts, who have been involved in a particular type of analysis for years and years. In Neuendorf&#x2019;s words (2002, p.&#x00A0;133): &#x201C;Three words describe good coder preparation: Train, train, and train.&#x201D; The more experience the coders have, the more likely it is that they are doing &#x2018;the same&#x2019; during the actual analysis.</p>
					<p>When deploying experts, it is important to check beforehand whether enough coders can be found with a comparable expertise who are unbiased in the sense that they are not aware of the research goal and hypotheses (cf. Artstein &amp; Poesio, 2008, pp.&#x00A0;574-575; Krippendorff, 2019, p.&#x00A0;274). Another issue is that intensive training of coders includes the risk of a coding bias: the trainees code the studied phenomenon the same way because they have learned to do this, which results in a high ICR. It is questionable, though, to what degree this affects the external validity of the study: is the research phenomenon still studied in the analysis, or did the coders learn some kind of superficial &#x2018;trick&#x2019;?</p>
				</sec>
			</sec>
		</sec>
		<sec sec-type="heading-1">
			<label>6</label>
			<title>Conclusions: Where do we go from here?</title>
			<p>For scholars of discourse studies using quantitative content analysis, issues of intercoder reliability are of the highest importance, both for practical reasons (how do we convince our peers that our studies are worthwhile despite our wild data yielding low ICR scores?) and for methodological reasons. In this paper, we have discussed how discourse data can differ in their degree of wildness and&#x00A0;&#x2013; with that&#x00A0;&#x2013; in their risk of a low ICR, how reliability can be at odds with validity, and what to pay attention to when assessing intercoder reliability statistically. We have also demonstrated that intercoder reliability issues are not insurmountable. Implementing the measures as discussed in the previous section will help improve the ICR scores of categorizing predefined units. Since assigning categories to discourse units is an essential part of the discourse coding process, we also believe that these measures will be beneficial to the coding tasks that combine unitizing a continuum and categorizing these units (for example Mathet et al., 2015). Nevertheless, we see important challenges for future work, amongst which big data and multimodal discourse.</p>
			<sec sec-type="heading-2">
				<label>6.1</label>
				<title>Big data</title>
				<p>A first issue is how to maintain insightful analyses when confronted with big data. In present-day corpus-based analyses, the availability of large quantities of discourse data raises all sorts of interesting opportunities compared to small-scale analyses, but also many problems. On the plus side, we have the possibility to look at our phenomena of interest in large numbers of texts, consisting of a wide range of genres, which increases the richness of our analyses and the generalizability of the results. At the same time, the sheer amount of data forces us to complement our manual analyses with automatic procedures, which can lead to ill-informed decisions in comparison to human annotations.</p>
				<p>A good example of the problems that automatic analyses can yield is provided by Vis (2011). She wanted to distinguish between words from the journalist and words from quoted sources in a wide variety of news texts from the 1950s and the 2000s. To automate this identification, she used the strategy of searching for quotation marks as the indicator of quoted sources. Although efficient, it is also a very coarse measure for quoted discourse. It neglects all forms of indirect and free indirect speech and writing, and it relies on the systematization with which the journalists made use of quotation marks. Unsurprisingly, such an automated procedure forces the researcher to build in manual checks on the quality of the resulting analysis.</p>
				<p>A possible improvement is the use of machine learning. Automatic classification by machine learning can be of assistance in manual coding tasks. Van den Bosch et al. (2006) described the word class or part-of-speech annotation of 50 million words, in which an automatic tagger was applied as a first filtering step. Typical for machine learning, the tagger combined the classification of a word with a confidence score for each possible part-of-speech tag, and only the words that could be assigned to different part-of-speech tags and stayed under a selected confidence threshold were flagged for manual annotation. The other words were labelled with the automatically assigned tag. This way, the number of units to be annotated manually decreased; Van den Bosch et al. (2006) claim that 28&#x202F;% of the cases in which the automatic part-of-speech tagger was most confident could be safely ignored as correct, and that only 1&#x202F;% of errors were missed. In addition, the confidence scores for different categories, along with information about common mistakes, guided the human annotators in their decision, which increased the ICR. In a similar vein, Reijnierse et al. (in preparation) use machine learning to identify metaphor-related words in a corpus of more than 90,000 Dutch newspaper articles on COVID19. The sheer size of such corpora prevents manual annotation of them.</p>
				<p>It should be stressed that such a procedure is especially feasible for low-level linguistic tasks such as part-of-speech tagging, for which most systems offer a sufficiently high base-level of performance and a relatively high reliability of their certainty score (Van den Bosch et al., 2006). Higher-level tasks are more challenging for machine learning. Sarcasm, for example, cannot be inferred from the presence of specific words or word clusters alone. Many markers of sarcasm&#x00A0;&#x2013; such as intensifiers, exclamation points, smileys, quotation marks&#x00A0;&#x2013; are ambiguous or hard to implement in machine learning&#x00A0;&#x2013; such as a change of register. The context is crucial to determine whether &#x201C;<italic>The weather is fantastic! ;-)</italic>&#x201D; is meant sarcastically or literally, which makes manual coders indispensable to arrive at reliable categorizations (Kunneman et al., 2015).</p>
				<p>That said, recent developments in natural language processing under the header of large language models (more generally referred to as generative AI or foundation models), popularized by systems such as ChatGPT appear to expand the capabilities of previous-generation machine learning methods. With few-shot learning or fine tuning, transformer-based deep learning systems are reported to be successful in many NLP tasks (Vaswani et al., 2018). However, calibrating their self-reported confidence to actual accuracy remains an issue (OpenAI, 2023); moreover, it has been argued that closed and proprietary large language models such as GPT-4 should not be used due to their inherent bias, which is in turn due to the large amount of biased data in their training datasets (of which the details are also kept in the dark). Open academic alternatives are there and are becoming more prevalent and of increasing quality (e.g., the BLOOM<xref ref-type="fn" rid="FN000003"><sup>3</sup></xref> family of large language models).</p>
			</sec>
			<sec sec-type="heading-2">
				<label>6.2</label>
				<title>Multimodal discourse</title>
				<p>Another challenge for the near future is the annotation of multimodal discourse. Present-day discourse frequently combines different modes of communication: verbal and visual, static and dynamic. Consider a TV commercial, consisting of text as seen on screen, combined with a voice-over describing the quality of the product, a clip showing a sequence of events, plus a static depiction of the logo and the product at the end of the commercial. How do we analyze this combination of written language plus visuals plus spoken language plus the interactions between all of these features? We are dealing with the combination of codes that differ fundamentally, in that the verbal code is basically non-iconic, as opposed to the iconic nature of the visuals. Although the study of multimodal discourse is booming (for example Bateman, 2008; Bateman &amp; Hiippala, 2021; Bateman &amp; Wildfeuer, 2014; Mordecai, 2023), attention to the ICR of this multifaceted type of discourse is scarce. An interesting initiative from the Metaphor Lab Amsterdam is to use the existing knowledge of metaphor in verbal language to analyze visual metaphor in their subprojects VisMet (Visual Metaphor, <ext-link xlink:href="http://www.vismet.org/VisMet/">http://www.vismet.org/VisMet/</ext-link>) and CogVim (Cognitive Grounding of Visual Metaphor, <ext-link xlink:href="https://cogvim.wordpress.com/">https://cogvim.wordpress.com/</ext-link>). The VisMIP (Visual Metaphor Identification Procedure) seeks to identify the metaphorical elements and their relationships in a reliable way (Steen, 2018). Other initiatives are the work by Taboada and Habel (2013) on rhetorical relations in multimodal documents and by Br&#x00F4;ne and Oben (2015) on gesture annotation. Other than that, there is to our knowledge no methodological work that particularly addresses the intercoder reliability of coding dynamic visuals, let alone the interaction between visuals and verbals, although the development of annotation tools such as ELAN (Sloetjes, 2014) and the MAST (Cardoso &amp; Cohn, 2022) will undoubtedly contribute to such work.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>6.3</label>
				<title>Concluding remarks</title>
				<p>Naturally occurring discourse data are messy. And in some cases, we need to annotate large amounts of data, in a restricted amount of time. It is no wonder researchers engaged in the quantitative corpus analysis of natural discourse sometimes feel they are in one of Augeas&#x2019; stables, not cleaned in over thirty years. We hope that the suggestions made in this paper contribute to dealing with such messiness and help discourse analysts to tame their wild data.</p>
			</sec>
		</sec>
	</body>
	<back>
		<ack>
			<title>Acknowledgements</title>
			<p>We thank the Centre for Language Studies at Radboud University for letting us organize the workshop on intercoder reliability that inspired this paper. We also thank the reviewers for their valuable feedback.</p>
		</ack>
		<fn-group>
			<fn id="FN000001"><label>1</label><p>To apply these metrics, we recommend the <ext-link xlink:href="https://www.nltk.org/"><underline>NLTK</underline> toolkit</ext-link> in the framework of the Python programming language, the packages offered in the <ext-link xlink:href="https://rdrr.io/github/jasp-stats/Reliability/f/inst/help/raterAgreement.md"><underline>JASP</underline></ext-link> or <ext-link xlink:href="https://www.jamovi.org/"><underline>JAMOVI</underline></ext-link> environment and in <ext-link xlink:href="https://www.rdocumentation.org/packages/psych/versions/2.3.3/topics/cohen.kappa"><underline>R</underline></ext-link>, and the package written by Andrew Hayes in the framework of SPSS or SAS.</p></fn>
			<fn id="FN000002"><label>2</label><p>Interestingly, Bayerl and Paul (2011, p.&#x00A0;713) suggest that high agreement established by a few number of coders may be due to chance agreement, and consequently they recommend to use a large number of coders when possible: &#x201C;The higher the number of annotators who are able to agree, the less bias and distortion can be expected&#x201D;. This issue is resolved when using metrics incorporating chance agreement.</p></fn>
			<fn id="FN000003"><label>3</label><p><ext-link xlink:href="https://huggingface.co/docs/transformers/model_doc/bloom">https://huggingface.co/docs/transformers/model_doc/bloom</ext-link></p></fn>
		</fn-group>
		<ref-list>
			<title>References</title>
			<ref id="R000001">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Arts, A., Maes, A., Noordman, L.G.M., &#x0026;amp; Jansen, C.</suffix></string-name></person-group> (<date><year>2011</year></date>). <article-title>Overspecification in written instruction</article-title>. <source><italic>Linguistics</italic></source>, <italic>49</italic>(<issue>3</issue>), <page-range>555</page-range>-<page-range>574</page-range>. <ext-link xlink:href="https://doi.org/10.1515/ling.2011.017">https://doi.org/10.1515/ling.2011.017</ext-link>.</mixed-citation>
			</ref>
			<ref id="R000002">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Artstein, R., &#x0026;amp; Poesio, M.</suffix></string-name></person-group> (<date><year>2008</year></date>). <article-title>Inter-coder agreement for computational linguistics</article-title>. <source><italic>Computational Linguistics</italic></source>, <italic>34</italic>(<issue>4</issue>), <page-range>555</page-range>-<page-range>596</page-range>. <ext-link xlink:href="https://doi.org/10.1162/coli.07-034-R2">https://doi.org/10.1162/coli.07-034-R2</ext-link>.</mixed-citation>
			</ref>
			<ref id="R000003">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Bateman</surname>, <given-names>J.</given-names></string-name></person-group> (<date><year>2008</year></date>). <source><italic>Multimodality and Genre: A Foundation for the Systematic Analysis of Multimodal Documents</italic></source>. <publisher-name>Springer</publisher-name>.</mixed-citation>
			</ref>
			<ref id="R000004">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Bateman, J.A., &#x0026;amp; Hiippala, T.</suffix></string-name></person-group> (<date><year>2021</year></date>). <chapter-title>From data to patterns</chapter-title>. In <person-group person-group-type="editor"><string-name><given-names>J.</given-names> <surname>Pflaeging</surname></string-name></person-group>, <person-group person-group-type="editor"><string-name><given-names>J.</given-names> <surname>Wildfeuer</surname></string-name></person-group>, <person-group person-group-type="editor"><string-name><given-names>J.A.</given-names> <surname>Bateman</surname></string-name></person-group> (Eds.), <source><italic>Empirical Mulitmodality Research. Methods, Evaluations, Implications</italic></source> (pp. <page-range>65</page-range>-<page-range>90</page-range>). <publisher-name>De Gruyter</publisher-name>. <ext-link xlink:href="https://doi.org/10.1515/9783110725001-003">https://doi.org/10.1515/9783110725001-003</ext-link></mixed-citation>
			</ref>
			<ref id="R000005">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Bateman, J.A., &#x0026;amp; Wildfeuer, J.</suffix></string-name></person-group> (<date><year>2014</year></date>). <article-title>A multimodal discourse theory of visual narrative</article-title>. <source><italic>Journal of Pragmatics</italic></source>, <italic>74</italic>, <page-range>180</page-range>-<page-range>208</page-range>. <ext-link xlink:href="https://doi.org/10.1016/j.pragma.2014.10.001">https://doi.org/10.1016/j.pragma.2014.10.001</ext-link></mixed-citation>
			</ref>
			<ref id="R000006">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Bayerl, P.S., &#x0026;amp; Paul, K.I.</suffix></string-name></person-group> (<date><year>2011</year></date>). <article-title>What determines inter-coder agreement in manual annotations? A meta-analytic investigation</article-title>. <source><italic>Computational Linguistics</italic></source>, <italic>37</italic>(<issue>4</issue>), <page-range>699</page-range>-<page-range>725</page-range>. <ext-link xlink:href="https://doi.org/10.1162/COLI_a_00074">https://doi.org/10.1162/COLI_a_00074</ext-link></mixed-citation>
			</ref>
			<ref id="R000007">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Bolly, C., Crible, L., Degand, L., &#x0026;amp; Uygur, D.</suffix></string-name></person-group> (<date><year>2014</year></date>). <chapter-title>Towards a model for discourse marker annotation in spoken French: From potential to feature-based discourse markers</chapter-title>. <person-group person-group-type="editor"><string-name><suffix>In A. Sans&#x00F3; &#x0026;amp; C. Fedriani (Eds.</suffix></string-name></person-group>), <source><italic>Pragmatic markers, discourse markers and modal particles: What do we know and where do we go from here?</italic></source> (pp. <page-range>71</page-range>-<page-range>97</page-range>). Benjamins. <ext-link xlink:href="https://dial.uclouvain.be/pr/boreal/object/boreal:161997">https://dial.uclouvain.be/pr/boreal/object/boreal:161997</ext-link></mixed-citation>
			</ref>
			<ref id="R000008">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Br&#x00F4;ne, G., &#x0026;amp; Oben, B.</suffix></string-name></person-group> (<date><year>2015</year></date>). <article-title>InSight Interaction: A multimodal and multifocal dialogue corpus</article-title>. <source><italic>Language Resources and Evaluation</italic></source>, <italic>49</italic>(<issue>1</issue>), <page-range>195</page-range>-<page-range>214</page-range>. <ext-link xlink:href="https://doi.org/10.1007/s10579-014-9283-2">https://doi.org/10.1007/s10579-014-9283-2</ext-link></mixed-citation>
			</ref>
			<ref id="R000009">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Burgers, C., Konijn, E.A., &#x0026;amp; Steen, G.J.</suffix></string-name></person-group> (<date><year>2016</year></date>). <article-title>Figurative framing: Shaping public discourse through metaphor, hyperbole, and irony</article-title>. <source><italic>Communication Theory</italic></source>, <italic>26</italic>(<issue>4</issue>), <page-range>410</page-range>-<page-range>430</page-range>. <ext-link xlink:href="https://doi.org/10.1111/comt.12096">https://doi.org/10.1111/comt.12096</ext-link></mixed-citation>
			</ref>
			<ref id="R000010">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Burgers, C., Van Mulken, M., &#x0026;amp; Schellens, P.J.</suffix></string-name></person-group> (<date><year>2011</year></date>). <article-title>Finding irony: An introduction of the Verbal Irony Procedure (VIP</article-title>). <source><italic>Metaphor and Symbol</italic></source>, <italic>26</italic>(<issue>3</issue>), <page-range>186</page-range>-<page-range>205</page-range>. <ext-link xlink:href="https://doi.org/10.1080/10926488.2011.583194">https://doi.org/10.1080/10926488.2011.583194</ext-link></mixed-citation>
			</ref>
			<ref id="R000011">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Byrt, T., Bishop, J., &#x0026;amp; Carlin, J.B.</suffix></string-name></person-group> (<date><year>1993</year></date>). <article-title>Bias, prevalence and kappa</article-title>. <source><italic>Journal of Clinical Epidemiology</italic></source>, <italic>46</italic>(<issue>5</issue>), <page-range>423</page-range>-<page-range>429</page-range>. <ext-link xlink:href="https://doi.org/10.1016/0895-4356(93)90018-V">https://doi.org/10.1016/0895-4356(93)90018-V</ext-link></mixed-citation>
			</ref>
			<ref id="R000012">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Cardoso, B., &#x0026;amp; Cohn, N.</suffix></string-name></person-group> (<date><year>2022</year></date>). <source>The Multimodal Annotation Software Tool (MAST</source>). <source>Proceedings of the 13th Conference on Language Resources and Evaluation (LREC 2022</source>), <page-range>6822</page-range>-<page-range>6828</page-range>.</mixed-citation>
			</ref>
			<ref id="R000013">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Cohen</surname>, <given-names>J.</given-names></string-name></person-group> (<date><year>1960</year></date>). <article-title>A coefficient of agreement for nominal scales</article-title>. <source><italic>Educational and Psychological Measurement</italic></source>, <italic>20</italic>(<issue>1</issue>), <page-range>37</page-range>-<page-range>46</page-range>. <ext-link xlink:href="https://doi.org/10.1177/001316446002000104">https://doi.org/10.1177/001316446002000104</ext-link></mixed-citation>
			</ref>
			<ref id="R000014">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Cohen</surname>, <given-names>J.</given-names></string-name></person-group> (<date><year>1968</year></date>). <article-title>Weighted kappa: Nominal scale agreement provision for scaled disagreement or partial credit</article-title>. <source><italic>Psychological Bulletin</italic></source>, <italic>70</italic>, <page-range>213</page-range>-<page-range>220</page-range>. <ext-link xlink:href="https://doi.org/10.1037/h0026256">https://doi.org/10.1037/h0026256</ext-link></mixed-citation>
			</ref>
			<ref id="R000015">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>De Smedt, T., &#x0026;amp; Daelemans, W.</suffix></string-name></person-group> (<date><year>2012</year></date>). <article-title>Pattern for python</article-title>. <source><italic>The Journal of Machine Learning Research</italic></source>, <italic>13</italic>, <page-range>2063</page-range>-<page-range>2067</page-range>.</mixed-citation>
			</ref>
			<ref id="R000016">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Di Eugenio, B., &#x0026;amp; Glass, M.</suffix></string-name></person-group> (<date><year>2004</year></date>). <article-title>The Kappa statistic: A second look</article-title>. <source><italic>Computational Linguistics</italic></source>, <italic>30</italic>(<issue>1</issue>), <page-range>95</page-range>-<page-range>101</page-range>. <ext-link xlink:href="https://doi.org/10.1162/089120104773633402">https://doi.org/10.1162/089120104773633402</ext-link></mixed-citation>
			</ref>
			<ref id="R000017">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Elmes, D.G., Kantowitz, B.H., &#x0026;amp; Roediger, H.L.I.</suffix></string-name></person-group> (<date><year>2012</year></date>). <source><italic>Research Methods in Psychology</italic></source> (9th ed). <publisher-name>Wadsworth Cengage Learning</publisher-name>.</mixed-citation>
			</ref>
			<ref id="R000018">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Feinstein, A.R., &#x0026;amp; Cicchetti, D.V.</suffix></string-name></person-group> (<date><year>1990</year></date>). <article-title>High agreement but low Kappa: I. the problems of two paradoxes</article-title>. <source><italic>Journal of Clinical Epidemiology</italic></source>, <italic>43</italic>(<issue>6</issue>), <page-range>543</page-range>-<page-range>549</page-range>. <ext-link xlink:href="https://doi.org/10.1016/0895-4356(90)90158-L">https://doi.org/10.1016/0895-4356(90)90158-L</ext-link></mixed-citation>
			</ref>
			<ref id="R000019">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Fleiss</surname>, <given-names>J.L.</given-names></string-name></person-group> (<date><year>1971</year></date>). <article-title>Measuring nominal scale agreement among many raters</article-title>. <source><italic>Psychological Bulletin</italic></source>, <italic>76</italic>, <page-range>378</page-range>-<page-range>382</page-range>. <ext-link xlink:href="https://doi.org/10.1037/h0031619">https://doi.org/10.1037/h0031619</ext-link></mixed-citation>
			</ref>
			<ref id="R000020">
				<mixed-citation><person-group person-group-type="editor"><string-name><suffix>Forceville, C. (Charles), &#x0026;amp; Urios-Aparisi, E. (Eds.).</suffix></string-name></person-group> (<date><year>2009</year></date>). <source><italic>Multimodal Metaphor</italic></source>. <publisher-name>Mouton de Gruyter</publisher-name>.</mixed-citation>
			</ref>
			<ref id="R000021">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Fort, K., Nazarenko, A., &#x0026;amp; Rosset, S.</suffix></string-name></person-group> (<date><year>2012</year></date>). <source>Modeling the complexity of manual annotation tasks: A grid of analysis</source>. In <italic>Proceedings of the International Conference on Computational Linguistics (COLING 2012)</italic> (pp. <page-range>895</page-range>-<page-range>910</page-range>). <ext-link xlink:href="https://hal.science/hal-00769631">https://hal.science/hal-00769631</ext-link></mixed-citation>
			</ref>
			<ref id="R000022">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Grove, W.M., Andreasen, N.C., McDonald-Scott, P., Keller, M.B., &#x0026;amp; Shapiro, R.W.</suffix></string-name></person-group> (<date><year>1981</year></date>). <article-title>Reliability studies of psychiatric diagnosis: Theory and practice</article-title>. <source><italic>Archives of General Psychiatry</italic></source>, <italic>38</italic>(<issue>4</issue>), <page-range>408</page-range>-<page-range>413</page-range>. <ext-link xlink:href="https://doi.org/10.1001/archpsyc.1981.01780290042004">https://doi.org/10.1001/archpsyc.1981.01780290042004</ext-link></mixed-citation>
			</ref>
			<ref id="R000023">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Hancock, J.T., Curry, L.E., Goorha, S., &#x0026;amp; Woodworth, M.</suffix></string-name></person-group> (<date><year>2007</year></date>). <article-title>On lying and being lied to: A linguistic analysis of deception in computer-mediated communication</article-title>. <source><italic>Discourse Processes</italic></source>, <italic>45</italic>(<page-range>1</page-range>), <issue>1</issue>-<page-range>23</page-range>. <ext-link xlink:href="https://doi.org/10.1080/01638530701739181">https://doi.org/10.1080/01638530701739181</ext-link></mixed-citation>
			</ref>
			<ref id="R000024">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Hayes, A.F., &#x0026;amp; Krippendorff, K.</suffix></string-name></person-group> (<date><year>2007</year></date>). <article-title>Answering the call for a standard reliability measure for coding data</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <italic>1</italic>(<issue>1</issue>), <page-range>77</page-range>-<page-range>89</page-range>. <ext-link xlink:href="https://doi.org/10.1080/19312450709336664">https://doi.org/10.1080/19312450709336664</ext-link></mixed-citation>
			</ref>
			<ref id="R000025">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Hoek, J., Sanders, T., &#x0026;amp; Spooren, W.</suffix></string-name></person-group> (<date><year>2021</year></date>). <article-title>Automatic coherence analysis of Dutch: Testing the subjectivity hypothesis on a larger scale</article-title>. <source><italic>Corpora</italic></source>, <italic>16</italic>(<issue>1</issue>), <page-range>129</page-range>-<page-range>155</page-range>. <ext-link xlink:href="https://doi.org/10.3366/cor.2021.0211">https://doi.org/10.3366/cor.2021.0211</ext-link></mixed-citation>
			</ref>
			<ref id="R000026">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Jeni, L.A., Cohn, J.F., &#x0026;amp; De La Torre, F.</suffix></string-name></person-group> (<date><year>2013</year></date>). <source>Facing imbalanced data &#x2013; Recommendations for the use of performance metrics</source>. <source><italic>2013 Humaine Association Conference on Affective Computing and Intelligent Interaction</italic></source>, <page-range>245</page-range>-<page-range>251</page-range>. <ext-link xlink:href="https://doi.org/10.1109/ACII.2013.47">https://doi.org/10.1109/ACII.2013.47</ext-link></mixed-citation>
			</ref>
			<ref id="R000027">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Krippendorff</surname>, <given-names>K.</given-names></string-name></person-group> (<date><year>2011</year></date>). <article-title>Agreement and information in the reliability of coding</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <italic>5</italic>(<issue>2</issue>), <page-range>93</page-range>-<page-range>112</page-range>.</mixed-citation>
			</ref>
			<ref id="R000028">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Krippendorff</surname>, <given-names>K.</given-names></string-name></person-group> (<date><year>2019</year></date>). <source><italic>Content Analysis: An Introduction to its Methodology</italic></source> (4th ed.). <publisher-name>SAGE</publisher-name>.</mixed-citation>
			</ref>
			<ref id="R000029">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Krippendorff, K., Mathet, Y., Bouvry, S., &#x0026;amp; Widl&#x00F6;cher, A.</suffix></string-name></person-group> (<date><year>2016</year></date>). <article-title>On the reliability of unitizing textual continua: Further developments</article-title>. <source><italic>Quality &#x0026;amp; Quantity</italic></source>, <italic>50</italic>(<issue>6</issue>), <page-range>2347</page-range>-<page-range>2364</page-range>. <ext-link xlink:href="https://doi.org/10.1007/s11135-015-0266-1">https://doi.org/10.1007/s11135-015-0266-1</ext-link></mixed-citation>
			</ref>
			<ref id="R000030">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Kunneman, F., Liebrecht, C., Van Mulken, M., &#x0026;amp; Van den Bosch, A.</suffix></string-name></person-group> (<date><year>2015</year></date>). <article-title>Signaling sarcasm: From hyperbole to hashtag</article-title>. <source><italic>Information Processing &#x0026;amp; Management</italic></source>, <italic>51</italic>(<issue>4</issue>), <page-range>500</page-range>-<page-range>509</page-range>. <ext-link xlink:href="https://doi.org/10.1016/j.ipm.2014.07.006">https://doi.org/10.1016/j.ipm.2014.07.006</ext-link></mixed-citation>
			</ref>
			<ref id="R000031">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Liebrecht</surname>, <given-names>C.</given-names></string-name></person-group> (<date><year>2015</year></date>). <italic>Intens Krachtig. Stilistische Intensiveerders in Evaluatieve Teksten [Intensely Powerful. Stylistic Intensifiers in Evaluative Texts.]</italic>. [Doctoral dissertation, <publisher-name>Radboud Universiteit</publisher-name>]. <ext-link xlink:href="https://hdl.handle.net/2066/141116">https://hdl.handle.net/2066/141116</ext-link></mixed-citation>
			</ref>
			<ref id="R000032">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Linders</surname>, <given-names>Y.</given-names></string-name></person-group> (<date><year>2014</year></date>). <italic>Met Waardering Gelezen. Een Nieuw Analyse-instrument en een Kwantitatieve Analyse van Evaluaties in Nederlandse Literaire Dagbladkritiek, 1955-2005 [Read with Appreciation. A New Instrument of Analysis and a Quantitative Analysis of Evaluations in Literary Reviews in Dutch Daily Newspapers]</italic>. [Doctoral dissertation, Radboud Universiteit]. <ext-link xlink:href="https://hdl.handle.net/2066/131544">https://hdl.handle.net/2066/131544</ext-link></mixed-citation>
			</ref>
			<ref id="R000033">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Martin, L.J., Degand, L., &#x0026;amp; Simon, A.-C.</suffix></string-name></person-group> (<date><year>2014</year></date>). <article-title>Forme et fonction de la p&#x00E9;riph&#x00E9;rie gauche dans un corpus oral multigenres annot&#x00E9; [Form and function of the left periphery in an annotated multi-genre oral corpus</article-title>]. <source><italic>Corpus</italic></source>, <italic>13</italic>. <ext-link xlink:href="https://doi.org/10.4000/corpus.2509">https://doi.org/10.4000/corpus.2509</ext-link></mixed-citation>
			</ref>
			<ref id="R000034">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Mathet, Y., Widl&#x00F6;cher, A., Fort, K., Fran&#x00E7;ois, C., Galibert, O., Grouin, C., Kahn, J., Rosset, S., &#x0026;amp; Zweigenbaum, P.</suffix></string-name></person-group> (<date><year>2012</year></date>). <source><italic>Manual corpus annotation: Giving meaning to the evaluation metrics</italic>, <italic>809</italic></source>. <ext-link xlink:href="https://hal.science/hal-00769639">https://hal.science/hal-00769639</ext-link></mixed-citation>
			</ref>
			<ref id="R000035">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Mathet, Y., Widl&#x00F6;cher, A., &#x0026;amp; M&#x00E9;tivier, J.-P.</suffix></string-name></person-group> (<date><year>2015</year></date>). <article-title>The unified and holistic method Gamma (<styled-content xml:lang="el-Grek">&#x03B3;</styled-content>) for inter-annotator agreement measure and alignment</article-title>. <source><italic>Computational Linguistics</italic></source>, <italic>41</italic>(<issue>3</issue>), <page-range>437</page-range>-<page-range>479</page-range>.</mixed-citation>
			</ref>
			<ref id="R000036">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Mol, L., Krahmer, E., Maes, A., &#x0026;amp; Swerts, M.</suffix></string-name></person-group> (<date><year>2012</year></date>). <article-title>Adaptation in gesture: Converging hands or converging minds?</article-title> <source><italic>Journal of Memory and Language</italic></source>, <italic>66</italic>(<issue>1</issue>), <page-range>249</page-range>-<page-range>264</page-range>. <ext-link xlink:href="https://doi.org/10.1016/j.jml.2011.07.004">https://doi.org/10.1016/j.jml.2011.07.004</ext-link></mixed-citation>
			</ref>
			<ref id="R000037">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Mordecai</surname>, <given-names>C.</given-names></string-name></person-group> (<date><year>2023</year></date>). <article-title>#anxiety: A multimodal discourse analysis of narrations of anxiety on TikTok</article-title>. <source><italic>Computers and Composition</italic></source>, <italic>67</italic>, <page-range>102763</page-range>. <ext-link xlink:href="https://doi.org/10.1016/j.compcom.2023.102763">https://doi.org/10.1016/j.compcom.2023.102763</ext-link></mixed-citation>
			</ref>
			<ref id="R000038">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Moss</surname>, <given-names>P.A.</given-names></string-name></person-group> (<date><year>1994</year></date>). <article-title>Can there be validity without reliability?</article-title> <source><italic>Educational Researcher</italic></source>, <italic>23</italic>(<issue>2</issue>), <page-range>5</page-range>-<page-range>12</page-range>. <ext-link xlink:href="https://doi.org/10.3102/0013189X023002005">https://doi.org/10.3102/0013189X023002005</ext-link></mixed-citation>
			</ref>
			<ref id="R000039">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Neuendorf</surname>, <given-names>K.A.</given-names></string-name></person-group> (<date><year>2002</year></date>). <source><italic>The Content Analysis Guidebook</italic></source>. <publisher-name>SAGE</publisher-name>.</mixed-citation>
			</ref>
			<ref id="R000040">
				<mixed-citation><person-group person-group-type="author"><string-name><given-names>OpenAI</given-names></string-name></person-group>. (<date><year>2023</year></date>). <source><italic>GPT-4 Technical Report</italic></source>. <ext-link xlink:href="https://doi.org/10.48550/ARXIV.2303.08774">https://doi.org/10.48550/ARXIV.2303.08774</ext-link></mixed-citation>
			</ref>
			<ref id="R000041">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Pasma</surname>, <given-names>T.</given-names></string-name></person-group> (<date><year>2011</year></date>). <source><italic>Metaphor and register variation: The personalization of Dutch news discourse</italic></source> [Doctoral dissertation, <publisher-name>VU University</publisher-name>]. <ext-link xlink:href="https://research.vu.nl/en/publications/metaphor-and-register-variation-the-personalization-of-dutch-news">https://research.vu.nl/en/publications/metaphor-and-register-variation-the-personalization-of-dutch-news</ext-link></mixed-citation>
			</ref>
			<ref id="R000042">
				<mixed-citation><person-group person-group-type="editor"><string-name><surname>Pe&#x00F1;a</surname>, <given-names>E.D.</given-names></string-name></person-group> (<date><year>2007</year></date>). <article-title>Lost in translation: Methodological considerations in cross-cultural research</article-title>. <source><italic>Child Development</italic></source>, <italic>78</italic>(<issue>4</issue>), <page-range>1255</page-range>-<page-range>1264</page-range>. <ext-link xlink:href="https://doi.org/10.1111/j.1467-8624.2007.01064.x">https://doi.org/10.1111/j.1467-8624.2007.01064.x</ext-link></mixed-citation>
			</ref>
			<ref id="R000043">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Perreault, Jr., W.D., &#x0026;amp; Leigh, L.E.</suffix></string-name></person-group> (<date><year>1989</year></date>). <article-title>Reliability of nominal data based on qualitative judgments</article-title>. <source><italic>Journal of Marketing Research</italic></source>, <italic>26</italic>, <page-range>135</page-range>-<page-range>148</page-range>.</mixed-citation>
			</ref>
			<ref id="R000044">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Potter, W.J., &#x0026;amp; Levine-Donnerstein, D.</suffix></string-name></person-group> (<date><year>1999</year></date>). <article-title>Rethinking validity and reliability in content analysis</article-title>. <source><italic>Journal of Applied Communication Research</italic></source>, <italic>27</italic>(<issue>3</issue>), <page-range>258</page-range>-<page-range>284</page-range>. <ext-link xlink:href="https://doi.org/10.1080/00909889909365539">https://doi.org/10.1080/00909889909365539</ext-link></mixed-citation>
			</ref>
			<ref id="R000045">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Quiros-Ramirez, M.A., &#x0026;amp; Onisawa, T.</suffix></string-name></person-group> (<date><year>2015</year></date>). <article-title>Considering cross-cultural context in the automatic recognition of emotions</article-title>. <source><italic>International Journal of Machine Learning and Cybernetics</italic></source>, <italic>6</italic>(<issue>1</issue>), <page-range>119</page-range>-<page-range>127</page-range>. <ext-link xlink:href="https://doi.org/10.1007/s13042-013-0192-2">https://doi.org/10.1007/s13042-013-0192-2</ext-link></mixed-citation>
			</ref>
			<ref id="R000046">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Reijnierse, G., Grunwald, J., &#x0026;amp; Spooren, W.</suffix></string-name></person-group> (in preparation). <italic>MetRobbert: Automatic metaphor identification in Dutch</italic>.</mixed-citation>
			</ref>
			<ref id="R000047">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Scholman, M.C.J., Evers-Vermeul, J., &#x0026;amp; Sanders, T.J.M.</suffix></string-name></person-group> (<date><year>2016</year></date>). <article-title>Categories of coherence relations in discourse annotation</article-title>. <source><italic>Dialogue &#x0026;amp; Discourse</italic></source>, <italic>7</italic>(<page-range>2</page-range>), <issue>2</issue>. <ext-link xlink:href="https://doi.org/10.5087/dad.2016.201">https://doi.org/10.5087/dad.2016.201</ext-link></mixed-citation>
			</ref>
			<ref id="R000048">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Scott</surname>, <given-names>W.A.</given-names></string-name></person-group> (<date><year>1955</year></date>). <article-title>Reliability of content analysis: The case of nominal scale coding</article-title>. <source><italic>Public Opinion Quarterly</italic></source>, <italic>19</italic>(<issue>3</issue>), <page-range>321</page-range>-<page-range>325</page-range>. <ext-link xlink:href="https://doi.org/10.1086/266577">https://doi.org/10.1086/266577</ext-link></mixed-citation>
			</ref>
			<ref id="R000049">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Selvi</surname>, <given-names>A.F.</given-names></string-name></person-group> (<date><year>2020</year></date>). <chapter-title>Qualitative content analysis</chapter-title>. <person-group person-group-type="editor"><string-name><suffix>In H. Rose &#x0026;amp; J. McKinley (Eds.</suffix></string-name></person-group>), <source><italic>The Routledge Handbook of Research Methods in Applied Linguistics</italic></source> (pp. <page-range>440</page-range>-<page-range>452</page-range>). <publisher-name>Routledge</publisher-name>.</mixed-citation>
			</ref>
			<ref id="R000050">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Sloetjes</surname>, <given-names>H.</given-names></string-name></person-group> (<date><year>2014</year></date>). <chapter-title>ELAN: Multimedia Annotation Application</chapter-title>. <person-group person-group-type="editor"><string-name><suffix>In J. Durand, U. Gut, &#x0026;amp; G. Kristoffersen (Eds.</suffix></string-name></person-group>), <source><italic>The Oxford Handbook of Corpus Phonology</italic></source>. <publisher-name>Oxford University Press</publisher-name>. <ext-link xlink:href="https://doi.org/10.1093/oxfordhb/9780199571932.013.019">https://doi.org/10.1093/oxfordhb/9780199571932.013.019</ext-link></mixed-citation>
			</ref>
			<ref id="R000051">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Spooren, W., &#x0026;amp; Degand, L.</suffix></string-name></person-group> (<date><year>2010</year></date>). <source><italic>Coding coherence relations: Reliability and validity</italic></source>. <italic>6</italic>(<issue>2</issue>), <page-range>241</page-range>-<page-range>266</page-range>. <ext-link xlink:href="https://doi.org/10.1515/cllt.2010.009">https://doi.org/10.1515/cllt.2010.009</ext-link></mixed-citation>
			</ref>
			<ref id="R000052">
				<mixed-citation><person-group person-group-type="editor"><string-name><surname>Steen</surname>, <given-names>G.J.</given-names></string-name></person-group> (Ed.) (<date><year>2018</year></date>). <source><italic>Visual Metaphor: Structure and process</italic></source>. <publisher-name>John Benjamins</publisher-name>. <ext-link xlink:href="https://doi.org/10.1075/celcr.18">https://doi.org/10.1075/celcr.18</ext-link></mixed-citation>
			</ref>
			<ref id="R000053">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Suviranta, R., &#x0026;amp; Hiippala, T.</suffix></string-name></person-group> (<date><year>2022</year></date>). <source><italic>Commercial crowdsourcing in digital humanities: Digital Humanities</italic></source>, <page-range>576</page-range>-<page-range>578</page-range>. <ext-link xlink:href="https://dh2022.dhii.asia/dh2022bookofabsts.pdf">https://dh2022.dhii.asia/dh2022bookofabsts.pdf</ext-link></mixed-citation>
			</ref>
			<ref id="R000054">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Taboada, M., &#x0026;amp; Habel, C.</suffix></string-name></person-group> (<date><year>2013</year></date>). <article-title>Rhetorical relations in multimodal documents</article-title>. <source><italic>Discourse Studies</italic></source>, <italic>15</italic>(<issue>1</issue>), <page-range>65</page-range>-<page-range>89</page-range>.</mixed-citation>
			</ref>
			<ref id="R000055">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Trochim</surname>, <given-names>W.M.K.</given-names></string-name></person-group> (<date><year>2006</year></date>). <source><italic>Reliability &#x0026;amp; Validity</italic></source>. <ext-link xlink:href="https://conjointly.com/kb/reliability-and-validity/">https://conjointly.com/kb/reliability-and-validity/</ext-link></mixed-citation>
			</ref>
			<ref id="R000056">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Umesh, U.N., Peterson, R.A., &#x0026;amp; Sauber, M.H.</suffix></string-name></person-group> (<date><year>1989</year></date>). <article-title>Interjudge agreement and the maximum value of Kappa</article-title>. <source><italic>Educational and Psychological Measurement</italic></source>, <italic>49</italic>(<issue>4</issue>), <page-range>835</page-range>-<page-range>850</page-range>. <ext-link xlink:href="https://doi.org/10.1177/001316448904900407">https://doi.org/10.1177/001316448904900407</ext-link></mixed-citation>
			</ref>
			<ref id="R000057">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Van den Bergh, H., Van Es, A., &#x0026;amp; Spijker, S.</suffix></string-name></person-group> (<date><year>2011</year></date>). <article-title>Spelling op verschillende niveaus: Werkwoordspelling aan het eind van de basisschool en het einde van het voortgezet onderwijs [Spelling at different levels: Verb spelling at the end of primary education and at the end of secondary education</article-title>]. <source><italic>Levende Talen Tijdschrift</italic></source>, <italic>12</italic>(<issue>1</issue>), <page-range>3</page-range>-<page-range>14</page-range>.</mixed-citation>
			</ref>
			<ref id="R000058">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Van den Bosch, A., Schuurman, I., &#x0026;amp; Vandeghinste, V.</suffix></string-name></person-group> (<date><year>2006</year></date>). <source>Transferring PoS-tagging and lemmatization tools from spoken to written Dutch corpus development</source>. In <italic>Proceedings of the Fifth International Conference on Language Resources and Evaluation, (LREC-2006)</italic>.</mixed-citation>
			</ref>
			<ref id="R000059">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Van Enschot, R., &#x0026;amp; Donn&#x00E9;, L.</suffix></string-name></person-group> (<date><year>2013</year></date>). <chapter-title>Retorische vormen in gezondheidsvoorlichting [Rhetorical figures in health communication</chapter-title>]. <person-group person-group-type="editor"><string-name><suffix>In R.J.U. Boogaart &#x0026;amp; H. Jansen (Eds.</suffix></string-name></person-group>), <source><italic>Studies in Taalbeheersing 4</italic></source> (pp. <page-range>91</page-range>-<page-range>101</page-range>). <publisher-loc>Van Gorcum</publisher-loc>.</mixed-citation>
			</ref>
			<ref id="R000060">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Van Enschot, R., &#x0026;amp; Hoeken, H.</suffix></string-name></person-group> (<date><year>2015</year></date>). <article-title>The occurrence and effects of verbal and visual anchoring of tropes on the perceived comprehensibility and liking of TV commercials</article-title>. <source><italic>Journal of Advertising</italic></source>, <italic>44</italic>(<issue>1</issue>), <page-range>25</page-range>-<page-range>36</page-range>. <ext-link xlink:href="https://doi.org/10.1080/00913367.2014.933688">https://doi.org/10.1080/00913367.2014.933688</ext-link></mixed-citation>
			</ref>
			<ref id="R000061">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Van Mulken, M., &#x0026;amp; Schellens, P.J.</suffix></string-name></person-group> (<date><year>2012</year></date>). <article-title>Over loodzware bassen en wapperende broekspijpen &#x2013; Gebruik en perceptie van taalintensiverende stijlmiddelen [On weighty basses and fluttering pant legs. Use and perception of intensifying stylistic devices</article-title>]. <source><italic>Tijdschrift Voor Taalbeheersing</italic></source>, <italic>34</italic>(<issue>1</issue>), <page-range>26</page-range>-<page-range>53</page-range>. <ext-link xlink:href="https://doi.org/10.5117/TVT2012.1.OVER418">https://doi.org/10.5117/TVT2012.1.OVER418</ext-link></mixed-citation>
			</ref>
			<ref id="R000062">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Van</suffix> <surname>Rijsbergen</surname>, <given-names>C.J.</given-names></string-name></person-group> (<date><year>1979</year></date>). <source><italic>Information Retrieval</italic></source> (2nd ed). <publisher-loc>Butterworths</publisher-loc>.</mixed-citation>
			</ref>
			<ref id="R000063">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., &#x0026;amp; Polosukhin, I.</suffix></string-name></person-group> (<date><year>2018</year></date>). <source>Attention Is All You Need</source>. <person-group person-group-type="editor"><string-name><suffix>In U. von Luxburg, I. Guyon, S. Bengio, H. Wallach, R. Fergus, S.V.N. Vishwanathan, &#x0026;amp; R. Garnett (Red.</suffix></string-name></person-group>), <italic>Advances in neural information processing systems 30: 31st Annual Conference on Neural Information Processing Systems (NIPS 2017): Long Beach, California, USA, 4-9 December 2017</italic> (Vol. 30). Curran Associates, Inc. <ext-link xlink:href="https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf">https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf</ext-link></mixed-citation>
			</ref>
			<ref id="R000064">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Vis</surname>, <given-names>K.</given-names></string-name></person-group> (<date><year>2011</year></date>). <source><italic>Subjectivity in News Discourse: A Corpus Linguistic Analysis of Informalization</italic></source> [Doctoral dissertation, <publisher-name>Vrije Universiteit Amsterdam</publisher-name>]. <ext-link xlink:href="https://research.vu.nl/en/publications/subjectivity-in-news-discourse-a-corpus-linguistic-analysis-of-in">https://research.vu.nl/en/publications/subjectivity-in-news-discourse-a-corpus-linguistic-analysis-of-in</ext-link></mixed-citation>
			</ref>
			<ref id="R000065">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Wallace</surname>, <given-names>B.C.</given-names></string-name></person-group> (<date><year>2015</year></date>). <article-title>Computational irony: A survey and new perspectives</article-title>. <source><italic>Artificial Intelligence Review</italic></source>, <italic>43</italic>(<issue>4</issue>), <page-range>467</page-range>-<page-range>483</page-range>. <ext-link xlink:href="https://doi.org/10.1007/s10462-012-9392-5">https://doi.org/10.1007/s10462-012-9392-5</ext-link></mixed-citation>
			</ref>
			<ref id="R000066">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Webber, B., Prasad, R., Lee, A., &#x0026;amp; Joshi, A.</suffix></string-name></person-group> (<date><year>2019</year></date>). <source><italic>The Penn Discourse Treebank 3.0 Annotation Manual</italic></source>. <ext-link xlink:href="https://catalog.ldc.upenn.edu/docs/LDC2019T05/PDTB3-Annotation-Manual.pdf">https://catalog.ldc.upenn.edu/docs/LDC2019T05/PDTB3-Annotation-Manual.pdf</ext-link></mixed-citation>
			</ref>
			<ref id="R000067">
				<mixed-citation><person-group person-group-type="author"><string-name><suffix>Zufferey, S., &#x0026;amp; Degand, L.</suffix></string-name></person-group> (<date><year>2013</year></date>). <article-title>Annotating the meaning of discourse connectives in multilingual corpora</article-title>. <source><italic>Corpus Linguistics and Linguistic Theory</italic></source>, <italic>13</italic>(<issue>2</issue>), <page-range>399</page-range>-<page-range>422</page-range>. <ext-link xlink:href="https://doi.org/10.1515/cllt-2013-0022">https://doi.org/10.1515/cllt-2013-0022</ext-link></mixed-citation>
			</ref>
		</ref-list>
	</back>
</article>