<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "https://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article article-type="research-article" xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="eng" dtd-version="1.1">
	<front>
		<journal-meta>
			<journal-id journal-id-type="eissn">2211-7253</journal-id>
			<journal-id journal-id-type="publisher-id">DuJAL</journal-id>
			<journal-title-group>
				<journal-title>Dutch Journal of Applied Linguistics</journal-title>
			</journal-title-group>
			<issn publication-format="print">2211-7245</issn>
			<issn publication-format="online">2211-7253</issn>
			<publisher>
				<publisher-name>Openjournals.nl</publisher-name>
			</publisher>
		</journal-meta>
		<article-meta>
			<article-id pub-id-type="doi">10.51751/DuJAL9420</article-id>
			<title-group>
				<article-title>Speaker discrimination as a function of vowel realization: does focus affect perception?</article-title>
			</title-group>
			<contrib-group>
				<contrib contrib-type="author" corresp="yes">
					<name name-style="western">
						<surname>Heeren</surname>
						<given-names>Willemijn</given-names>
					</name>
					<email xlink:href="mailto:w.f.l.heeren@hum.leidenuniv.nl">w.f.l.heeren@hum.leidenuniv.nl</email>
					<xref ref-type="aff" rid="AFF000001"/>
					<role>conceptualization, methodology, formal analysis, writing&#x00A0;&#x2013; original draft, writing&#x00A0;&#x2013; review &#x0026; editing, funding acquisition</role>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Voeten</surname>
						<given-names>Cesko</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000002"/>
					<xref ref-type="aff" rid="AFF000003"/>
					<role>methodology, formal analysis, writing&#x00A0;&#x2013; original draft, writing&#x00A0;&#x2013; review &#x0026; editing, visualization</role>
				</contrib>
				<contrib contrib-type="author">
					<name name-style="western">
						<surname>Marks</surname>
						<given-names>Tessi</given-names>
					</name>
					<xref ref-type="aff" rid="AFF000001"/>
					<role>conceptualization, investigation, writing&#x00A0;&#x2013; original draft, writing&#x00A0;&#x2013; review &#x0026; editing</role>
				</contrib>
				<aff id="AFF000001">
					<label>1</label>
					<institution-wrap>
						<institution>Leiden University</institution>
					</institution-wrap>
				</aff>
				<aff id="AFF000002">
					<label>2</label>
					<institution-wrap>
						<institution>Fryske Akademy</institution>
					</institution-wrap>
				</aff>
				<aff id="AFF000003">
					<label>3</label>
					<institution-wrap>
						<institution>Utrecht University</institution>
					</institution-wrap>
				</aff>
			</contrib-group>
			<pub-date publication-format="online">
				<day>23</day>
				<month>05</month>
				<year>2022</year>
			</pub-date>
			<volume content-type="number">11</volume>
			<volume content-type="year">2022</volume>
			<fpage specific-use="PDF">1</fpage>
			<lpage>20</lpage>
			<history>
				<date date-type="received">
					<day>02</day>
					<month>03</month>
					<year>2021</year>
				</date>
				<date date-type="accepted">
					<day>13</day>
					<month>03</month>
					<year>2022</year>
				</date>
				<date date-type="pub">
					<day>23</day>
					<month>05</month>
					<year>2022</year>
				</date>
			</history>
			<permissions>
				<copyright-statement>Copyright 2022 by the author(s).</copyright-statement>
				<copyright-year>2022</copyright-year>
				<copyright-holder>the author(s)</copyright-holder>
				<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/" xlink:title="CC BY">
					<license-p>This is an open access article distributed under the terms of the CC BY 4.0 license.</license-p>
				</license>
			</permissions>
			<self-uri content-type="PDF" xlink:href="DuJAL9420_text.pdf"/>
			<abstract>
				<title>Abstract</title>
				<p>The acoustic-phonetic characteristics of speech sounds are influenced by their linguistic position in an utterance. Because of acoustic-phonetic differences between different speech sounds, sounds vary in the amount of speaker information they contain. However, do spectral and durational differences between various realizations of the same sound that were sampled from different linguistic positions also impact speaker information? We investigated speaker discrimination in [&#x2212;focus] versus [+focus] word realizations. Twenty-one Dutch listeners participated in a same-different speaker discrimination task, using stimuli varying in focus, vowel ([a&#x02D0;], [u]), and word context ([&#x0266;_k], [v_t]), spoken by 11 different speakers. Results show that an effect of focus on speaker-dependent information was present, but limited to words containing [u]. Moreover, performance on [u] words was influenced by (interactions of) word context and trial type (same- vs. different-speaker). Context-dependent changes in a speech sound&#x2019;s acoustics may affect its speaker-dependent information, albeit under specific conditions only.</p>
			</abstract>
			<kwd-group kwd-group-type="uncontrolled">
				<title>Keywords</title>
				<kwd>speaker discrimination</kwd>
				<kwd>focus</kwd>
				<kwd>vowel quality</kwd>
				<kwd>speech perception</kwd>
				<kwd>Dutch</kwd>
			</kwd-group>
			<funding-group>
				<funding-statement>This research was supported by a VIDI grant (276-75-010) from the Dutch Research Council.</funding-statement>
			</funding-group>
			<custom-meta-group>
				<custom-meta>
					<meta-name>Conflicting interests</meta-name>
					<meta-value>The authors have declared that conflicts of interest did not exist.</meta-value>
				</custom-meta>
				<custom-meta>
					<meta-name>Supporting information</meta-name>
					<meta-value>None</meta-value>
				</custom-meta>
				<custom-meta>
					<meta-name>Appendices</meta-name>
					<meta-value>Appendix A: Statement-question pairs used to elicit [&#x00B1;focus] versions of the target words; Appendix B: Visualization of the two significant clusters for the by-trial random slope for the interaction</meta-value>
				</custom-meta>
			</custom-meta-group>
		</article-meta>
	</front>
	<body>
		<sec sec-type="heading-1">
			<label>1</label>
			<title>Introduction</title>
			<p>Different speech sounds vary in the amount of acoustic speaker information they carry (Van den Heuvel, 1996; Kavanagh, 2012); vowels tend to contain more acoustic information on the speaker than consonants do, and within the class of consonants, nasals (e.g. [n, m]) may carry more speaker information than fricatives (e.g. [f, z]) or stops (e.g. [t, b]). Also in speech perception, the speech sounds that make up an utterance influence how well a listener can discriminate speakers (Andics et al., 2007) or identify them (Amino &#x0026; Arai, 2007). Much acoustic-phonetic research, however, has shown that the realization of one particular speech sound in one particular phonetic context varies as a function of linguistic position. Examples from Dutch are how the realization of a speech sound varies with the presence/absence of stress (Sluijter &#x0026; Van Heuven, 1996), or with word class (Van Bergem, 1993). For instance, in a stressed versus unstressed position, the syllable &#x2018;kom&#x2019; in &#x2018;&#x02C8;kom&#x00B7;ma&#x2019; (<italic>comma</italic>) as opposed to that in &#x2018;kom&#x00B7;&#x02C8;pas&#x2019; (<italic>compass</italic>), has a longer duration and less vowel reduction (Van Bergem, 1995, p.&#x00A0;28). Such variation in a speech sound&#x2019;s acoustics might in turn affect the speaker information carried by the sound. Whereas earlier work has compared the speaker information carried by different speech sounds, the main research question in the current investigation is if speaker discrimination performance on the <italic>same</italic> speech sound depends on whether tokens of the speech sound are sampled from focused versus non-focused contexts. Focus may be used to indicate new or contrastive information, and this is accompanied by acoustic changes (e.g., Hanssen et al., 2008; Sluijter &#x0026; Van Heuven, 1996).</p>
			<p>In the rest of this introductory section, speaker information in the acoustics of speech sounds (henceforth also referred to as <italic>segments</italic>) will be explained further in 1.1. This is followed by a discussion of prior research on how speaker discrimination may depend on stimulus composition in 1.2. The research questions and hypotheses are formulated in section 1.3.</p>
			<sec sec-type="heading-2">
				<label>1.1</label>
				<title>Acoustic-phonetic speaker-dependent information in segments</title>
				<p>In theory, an acoustic-phonetic speech feature (such as duration, fundamental frequency or formant frequency) would be highly speaker-specific when individual speakers show little variation in that feature when producing different tokens, i.e. there is little within-speaker variation, whereas different speakers would produce tokens that are very different with respect to the feature, i.e. there is large between-speaker variation. In practice, the speaker-specificity or speaker-discriminatory potential of individual speech features tends to be low to moderate. But crucially, both within-speaker and between-speaker variation of particular features differ in magnitudes across linguistic contexts (e.g., McDougall, 2006; Smorenburg &#x0026; Heeren, 2020): in some situations speakers show more variation whereas in others they show less. This creates the option of particular linguistic contexts yielding higher speaker-specificity than others.</p>
				<p>Most research on the speaker-specificity of acoustic-phonetic information has investigated how much speaker information is carried by individual speech sounds, and by vowels in particular. An early comparative investigation into differences in speaker-dependent acoustics for various speech sounds in Dutch was done by Van den Heuvel (1996). The segments included in that study were the vowels [i, a, u] and consonants [p, t, k, d, s, m, n, r]. Analyses of productions gathered under controlled circumstances, i.e. read pseudo-words, showed that the vowel [a] was the most speaker-dependent segment and that plosives, such as [p, d] contained the least speaker information. The speaker-discriminating potential of a large set of German phonemes was studied by Schindler and Draxler (2013), who found that some consonants, that is [s, n, m, f], contained more speaker-dependent information than most of the vowels. As in Dutch (see Van den Heuvel, 1996), the vowel [a&#x02D0;] was found to contain more speaker-dependent information than [i, u] (Schindler &#x0026; Draxler, 2013). In Czech, the vowel [i&#x02D0;] was found to pattern more with [a&#x02D0;] than [u] (Fejlov&#x00E1; et al., 2013). Finally, various diphthongs were also found to differ in the speaker-specific information they carry (e.g. Morrison, 2009); the reason may be that, in addition to differences in vowel quality, diphthongs vary in the amount and direction of inherent spectral change. Taken together, the results on vowels suggest that in addition to vowel quality, inherent spectral change and vowel duration may contribute speaker-dependent acoustics. For consonants, the longer segments in which specific spectral characteristics (resonances) can be found seem to carry most speaker information.</p>
				<p>There is some evidence that the linguistic context from which a speech sound is sampled influences the amount of within- and between-speaker variation it contains. He and Dellwo (2017) found more between-speaker variability in mouth-closing than mouth-opening gestures in read speech. This was explained as the second half of a syllable having lower articulatory demands relative to the first half, thus allowing for more articulatory freedom in individual speakers. In a follow-up investigation, He, Zhang and Dellwo (2019) demonstrated that also the first formant, which reflects the degree of mouth opening during vowel articulation, shows more between-speaker variability over the second half of a syllable than the first. In a study investigating speaker-dependent information in Dutch fricatives [s] and [x], Smorenburg and Heeren (2020) found that coda fricatives showed more between-speaker variation than onset fricatives, whereas the within-speaker variation showed a change in the opposite direction. Speaker classification scores were slightly better in codas than onsets for Dutch [x], whereas for [s] no positional difference in accuracy was found. A later study on speaker-dependent information in Dutch nasals [n] and [m] showed higher speaker classification scores for [n] in codas than onsets, but an effect in the opposite direction for [m] (Smorenburg &#x0026; Heeren, 2021). This difference in classification was explained by effects of the phonetic context on speech sound acoustics; context effects were larger in onset [m]s and coda [n]s. Hence, in addition to an effect of articulatory freedom depending on syllable position, the speech sound&#x2019;s susceptibility to co-articulatory influences may also affect how much speaker information can be carried by a speech sound.</p>
				<p>As was mentioned in section 1 (under Introduction), the acoustics of a particular speech sound are also influenced by higher-level linguistic factors, such as the presence/absence of stress on the syllable containing the speech sound (Sluijter &#x0026; Van Heuven, 1996) or the word class that its carrier belongs to (Van Bergem, 1993; Heeren, 2020). Another construct that is known to affect the realization of utterances is focus, which for instance expresses which part of a sentence is new (narrow focus) or contrastive (contrastive focus). In Information Structure theory (Chafe, 1976), focus has been defined as indicating &#x2018;the presence of alternatives that are relevant for the interpretation of linguistic expressions&#x2019; (Krifka, 2007, p.&#x00A0;18). Focus may be marked in different ways; several experiments on Dutch word realization demonstrated that segmental and syllabic acoustic measures may vary with focus condition. For instance, in comparison with broad focus, narrow and contrastive focus were found to affect segmental duration and fundamental frequency contours (e.g. Hanssen et al., 2008, see also Chen, 2009).</p>
				<p>In sum, a speech sound&#x2019;s spectral and durational characteristics may change with its position in an utterance. It has been demonstrated that between-sound differences in e.g. spectral composition and duration affect speaker discrimination (see section 1.2. for more details). The question then is if the spectral and durational changes found <italic>within</italic> the same sound as a function of its linguistic context would impact speaker discrimination.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>1.2</label>
				<title>Speaker discrimination by linguistic content and context</title>
				<p>In listeners, short words of different segmental composition elicit differential speaker discrimination performance. In Andics et al. (2007), native Dutch listeners heard CVC words and gave same-different speaker decisions for subsequent tokens that were auditorily presented. The CVC&#x202F;s were made up of onset [m] or [l], nucleus [&#x03B5;] or [&#x0254;], and coda [s] or [t], thus yielding eight different Dutch words. Results showed that [m], [&#x03B5;] and [s] yielded better discrimination performance than their positional counterparts [l], [&#x0254;] and [t]. In a trained-to-familiar speaker recognition experiment, Drozdova et al. (2017) found that listener performance was positively affected by the presence of vowels and nasals. These findings roughly correspond to findings from speech acoustics on which speech sounds contain most speaker-dependent information (Van den Heuvel, 1996; Kavanagh, 2012; Schindler &#x0026; Draxler, 2013, and discussion in section 1.1).</p>
				<p>Consistent with Fant&#x2019;s source-filter model of speech production (Fant, 1960), Baumann and Belin (2010) found that the two principal components explaining unfamiliar voice discrimination on sustained vowel data were based on the vocal source on the one hand, and the vocal filter on the other. This suggests that voiced speech sounds have an advantage over unvoiced speech sounds in talker perception, because the former benefit from both information sources. This is supported by results from Orchard and Yarmey (1995), who showed that speaker identification in a whispered voice line-up after having heard a whispering &#x2018;perpetrator&#x2019; was worse than identification in an all-normal voice condition. They furthermore found that samples of longer duration resulted in better performance than shorter samples (see also Cook &#x0026; Wilding, 1997). Earlier, Bricker and Pruzansky (1966) showed that not only duration <italic>per se</italic>, but also the segmental content within the speech fragment affected listener performance; better identification was obtained with more phonemes in a same-duration sample. The importance of measures of acoustic variation was recently also reported by Lee, Keating and Kreiman (2019), when investigating the acoustic dimensions that explain within-speaker and between-speaker variation.</p>
				<p>The task of speaker discrimination not only entails perceiving differences between speakers, but also perceiving that certain differences belong to the same speaker (assuming that non-identical stimulus pairs are presented). It has been shown that such within-speaker variability is judged differently for familiar than unfamiliar voices (Lavan et al., 2019); given 30 speech samples produced by two voices (15 fragments each) listeners who were familiar with the voices clustered the samples into fewer speaker identities than listeners who were unfamiliar with the voices. Therefore, both within-speaker and between-speaker variation of various acoustic parameters seemingly contribute to the processing of speaker information.</p>
				<p>In addition to effects of segmental differences and acoustic variation on the processing of speaker information, earlier work has shown effects of higher levels of linguistic information. At the semantic level, for instance, Van Berkum et al. (2008) demonstrated that indexical information in a voice rapidly influences the semantic processing of what the voice is saying: the mentioning of an alcoholic beverage in a child&#x2019;s voice elicited a different ERP response than in an adult voice. Also, the linguistic relationship between words influences speaker discrimination performance: using a same-different speaker discrimination task, Narayan, Mak and Bialystok (2017) showed that listeners had a tendency to assign linguistically-related word combinations to the same speaker (e.g. words with the same phonological rhyme &#x2018;bay-day&#x2019; as opposed to unrelated words &#x2018;day-bee&#x2019;).</p>
				<p>Earlier perception studies have thus demonstrated a connection between speaker discrimination and higher-level linguistic information. The current study investigated how the presence versus absence of focus, which would alter the precise phonetic-acoustic content of the same speech sound, affects speaker discrimination. This investigation of sub-segmental differences thus adds to the literature on voice perception, which has so far shown that segments vary in speaker-dependent information, and that higher levels of linguistics may influence voice perception.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>1.3</label>
				<title>Research questions and hypotheses</title>
				<p>This study in the first place investigated if speaker discrimination performance is affected by the presence versus absence of focus on a word. This was done using a same-different perception task, in which listeners judged whether subsequent words were produced by the same or by different speakers, and word stimuli were sampled from focused and non-focused sentence positions (further details in 2.2). In order to include some variation in word stimuli, two vowel nuclei were chosen from the three Dutch corner vowels included in Van den Heuvel (1996), namely those vowels that differed most in acoustic speaker information. This study thus also investigated if the difference in speaker-specific information between the Dutch corner vowels [a&#x02D0;] and [u] that is found in acoustics is also observed in perception.</p>
				<p>We hypothesized that speaker discrimination is more accurate on [a&#x02D0;] than [u], following the acoustic literature. Differences in both spectral and durational information between the vowels may contribute to this effect. As for the effect of focus, the main topic of this study, the prediction is less straightforward. In the comparison of unfamiliar voices there is an important role for low-level acoustic information (cf. Stevenage, 2018). The comparison of acoustic information between two speech samples leads to a same- or different-voice decision when the listener considers if the perceived variation between two non-identical tokens falls under within-speaker or between-speaker variation. From the literature it is known that listeners tend to underestimate within-speaker variation in unfamiliar speakers (Lavan et al., 2019). Therefore, the closer two realizations are acoustically, the more likely they are to be judged as &#x2018;same-speaker&#x2019;. On the one hand, it can be argued that the relatively precise articulation in a [+focus] word may be more comparable from one token to the next than in a [&#x2212;focus] condition. Moreover, [+focus] words are expected to be longer, thus giving them a perceptual advantage. On the other hand, with an expected lower occurrence frequency of these canonical forms relative to unfocused realizations in everyday speech, it may also be the case that less articulatory routine results in more within-speaker variation in [+focus] than in [&#x2212;focus] forms. Moreover, between-speaker variation has been found to be higher in locations where articulatory demands are lower, such as in coda consonants and closing syllable gestures (e.g., He &#x0026; Dellwo, 2017; Smorenburg &#x0026; Heeren, 2020). In the current investigation the [&#x2212;focus] words are therefore predicted to show larger between-speaker variation. Note that only in combination with smaller within-speaker variation, higher speaker-specificity is expected.</p>
				<p>We assessed if effects of vowel quality and focus would hold across word contexts by including two carrier words. The word contexts [&#x0266;_k] and [v_t] were selected so that the speaker information contained by the different onset and offset consonants was estimated to be low and roughly comparable. Schindler and Draxler (2013) showed that the ratio of between-to-within speaker variation was quite low for [v] and [h] sampled from spontaneous German speech, with perhaps a small advantage for [v]. Moreover, according to Van den Heuvel (1996) Dutch stop consonants are expected to contribute relatively low speaker information. Moreover, these word contexts gave target word frequencies falling in the mid-frequency range (<italic>haak</italic>/<italic>hoek</italic>, <italic>vaat</italic>/<italic>voet</italic>); word frequency is known to influence articulation (Bell et al., 2009) as high-frequency words are more subject to reduction than low-frequency words.</p>
				<p>The speaker discrimination results are accompanied by an acoustic analysis, including within- and between-speaker variation, of how speech sounds differ between conditions. Underlying this investigation is the assumption that speech sounds&#x2019; acoustics vary with focus condition (Eefting, 1991; Van Heuven, 1997), but to ensure that this was also the case for the stimuli presented to our listeners, the acoustic analysis was carried out.</p>
			</sec>
		</sec>
		<sec sec-type="heading-1">
			<label>2</label>
			<title>Method</title>
			<p>Following Andics et al. (2007), a same-different forced-choice one-back task was used for the speaker discrimination perception task. This means that listeners were presented with a series of stimuli and decided if the speaker of the current stimulus is the same as or different from the speaker of the previous stimulus.</p>
			<sec sec-type="heading-2">
				<label>2.1</label>
				<title>Participants</title>
				<p>Twenty-one Dutch listeners without self-reported hearing problems volunteered to take part in this perception study (12 females, 9 males). Their mean age was 23 years (SD = 1.5 years). All participants beforehand gave their informed consent for taking part in the study, and afterwards received a modest thank-you gift.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>2.2</label>
				<title>Stimuli</title>
				<p>Two Dutch minimal word pairs were used as stimulus contexts for the vowels [a&#x02D0;] and [u]: <italic>haak</italic>/<italic>hoek</italic> ([&#x0266;a&#x02D0;k]/[&#x0266;uk], &#x2018;hook&#x2019;/&#x2018;corner&#x2019;) and <italic>vaat</italic>/<italic>voet</italic> ([va&#x02D0;t]/[vut], &#x2018;dishes&#x2019;/&#x2018;foot&#x2019;). These words, with word frequencies in the mid-frequency range, were selected using the SUBTLEX-NL corpus (Keuleers et al., 2010): <italic>haak</italic> (12.6/million), <italic>hoek</italic> (49.4/million), <italic>vaat</italic> (1.4/million), <italic>voet</italic> (50.8/million).</p>
				<p>These four words were each recorded in a sentence context to evoke realizations with and without focus. The target sentences were answers to statement-question pairs that speakers saw on a computer screen, and speakers were instructed to produce answers in the form of full sentences. Two examples of the statement-question pairs and their intended answers are given here (see the appendix for a full list), where the first pair was intended to elicit [&#x2212;focus] tokens of the target word, and the second pair to elicit [+focus] tokens:</p>
				<def-list>
					<def-item>
						<term id="TRM000001">Prompt 1:</term>
						<def>
							<p>Hij zet zijn vaat in de wasbak. Waar zet hij zijn vaat?</p>
							<p><italic>He puts his dishes in the sink. Where does he put his dishes?</italic></p>
						</def>
					</def-item>
					<def-item>
						<term id="TRM000002">Answer 1:</term>
						<def>
							<p>Hij zet zijn vaat in de <bold>wasbak</bold>.</p>
							<p><italic>He puts his dishes in the sink</italic>.</p>
						</def>
					</def-item>
					<def-item>
						<term id="TRM000003">Prompt 2:</term>
						<def>
							<p>Hij zet zijn vaat in de wasbak. Wat zet hij in de wasbak?</p>
							<p><italic>He puts his dishes in the sink. What does he put in the sink?</italic></p>
						</def>
					</def-item>
					<def-item>
						<term id="TRM000004">Answer 2:</term>
						<def>
							<p>Hij zet zijn <bold>vaat</bold> in de wasbak.</p>
							<p><italic>He puts his dishes in the sink</italic>.</p>
						</def>
					</def-item>
				</def-list>
				<p>Eleven male speakers of Standard Dutch, aged 20 to 26 years, were recruited for the recordings. The stimuli were recorded in a sound-attenuated booth at the Leiden University Centre for Linguistics, using Praat software (Boersma &#x0026; Weenink, 2018), a Sennheiser MKH&#x00A0;416T microphone, and a FocusRite Scarlet 2i4 sound card. Recordings were saved as mono wave files (22,050 Hz, 16 bits). In an information sheet, speakers were instructed to produce the answer to a question about a short statement (see example above). Both the statement and question were shown orthographically on a computer screen in the recording booth. Before participation, speakers gave their informed consent.</p>
				<p>Each speaker produced each target word-focus combination six times, so that a sufficient number of tokens would be available for the perception experiment in which five repetitions of each would be needed. Tokens with the highest signal intensity relative to the background noise were kept for further processing. Target words were cut from the carrier sentences, resulting in 440 stimuli (11 speakers &#x00D7; 5 repetitions &#x00D7; 2 focus conditions &#x00D7; 2 vowel nuclei ([a&#x02D0;], [u]) &#x00D7; 2 word contexts ([v_t], [&#x0266;_k]). Stimulus intensities were all set to 65 dB SPL.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>2.3</label>
				<title>Stimulus acoustics and their statistical assessment</title>
				<p>As a first step, we evaluated statistically whether stimulus acoustics varied by focus condition and by vowel, as they are commonly assumed to do. The results are presented in section 3.1. In addition to stimulus duration, the mean fundamental frequency (F0) was taken as a measure of vocal source information, and as vocal filter parameters the first and second formants (F1 and F2) were extracted. Using Praat (Boersma &#x0026; Weenink, 2018), F0 was measured over the full duration of a stimulus using an autocorrelation method, and formants were measured using the Burg method at the point in time where vowel intensity was maximal (window size = 25 ms).</p>
				<p>Acoustic effects of the fixed factors Vowel, Focus and Word Context were evaluated in linear mixed-effects models (with &#x03B1; Bonferroni-corrected to .05/4 = .0125, given four acoustic measures). Factor levels [a&#x02D0;], [&#x2212;focus] and [&#x0266;_k] were used default levels, with predictions for acoustic differences between factor levels being directional, e.g. duration is expected to shorten from [a&#x02D0;] to [u], and to lengthen from [&#x2212;focus] to [+focus]. Using function buildmer from R package buildmer (Voeten, 2020), the maximally-converging models were obtained following a stepwise forward procedure. Duration was log<sub>10</sub>-transformed and F0, F1 and F2 were transformed to the Bark frequency scale before modelling.</p>
				<p>Both within-speaker and between-speaker variances were determined by vowel, by focus condition and by word context. The ratio of between- to within-speaker variances is called the &#x2018;speaker-specificity index&#x2019; (SSI, Van den Heuvel, 1996, p.&#x00A0;53), which was computed per acoustic parameter; the larger the variation between speakers relative to that within speakers, the higher the SSI, and the better speakers can presumably be separated. Because this is a descriptive view on the data, the measurements in Hertz and milliseconds were used for the sake of interpretability.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>2.4</label>
				<title>Design and procedure of the perception task</title>
				<p>For each of the four words included (<italic>haak, hoek, vaat, voet</italic>), a pseudo-random presentation list was made, in which each different-speaker pair occurred once (11&#x202F;&#x00D7;&#x202F;10/2) and each same-speaker pair occurred five times (11&#x202F;&#x00D7;&#x202F;5). In this way the same number (i.e. 55) of different- and same-speaker trials were included per word. The latter trial type made use of the different recordings from the same speaker, so that no identical recordings were used in a comparison. During the full experiment, each individual token was used twice, and different tokens by the same speaker were presented not more than three times in a row.</p>
				<p>Stimulus lists were distributed over 24 presentation blocks, containing one of eight vowel+word context+focus combinations each (3 blocks/combination). Per word context, there were 110 pairs for comparison, divided into blocks of 37, 37 and 36 trials each. A block lasted for about 1.5 minutes and the order of the blocks was randomized per listener. Across listeners 18,480 responses were collected. In this one-back discrimination task, not all speaker pairs occurred in both orders equally frequently. In the statistical analysis this is controlled for by the inclusion of random effects in the modelling.</p>
				<p>To each token&#x2019;s onset and offset a 5-ms fade-in or fade-out was applied in order to prevent clicks at trial onset or offset. Between subsequent stimuli, i.e. during the listener&#x2019;s response time, pink noise at an intensity of 50 dB was played. The next stimulus started 2,400 ms after the onset of the previous one.</p>
				<p>The perception experiment was run in a sound-attenuated booth at the phonetics laboratory of the Leiden University Centre for Linguistics, using E-Prime (Psychology Software Tools, 2012). Stimuli were presented at a standardized, comfortable listening level, over Beyerdynamic DT&#x00A0;770 PRO headphones.</p>
				<p>Listeners were instructed to carefully listen to the subsequent tokens, and to respond, after every token (but the first), whether the speaker of the latter token was the same as that of the former token or not. Responses were given by pressing one of two buttons on a QWERTY keyboard, &#x2018;X&#x2019; or &#x2018;N&#x2019;, one for &#x2018;same&#x2019; and the other for &#x2018;different&#x2019; speaker. The response buttons were counterbalanced across listeners.</p>
				<p>Before the actual experiment started, listeners completed a short practice round including tokens of the word <italic>vis</italic> (&#x2018;fish&#x2019;) to get used to the task. Including instruction, practice and breaks, the experiment lasted for about 45 minutes.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>2.5</label>
				<title>Statistical analysis of the perception data</title>
				<p>Initial examination of the mean correct responses per speaker showed that none of the eleven voices were especially hard or easy for the listeners. Hence, all speakers were included in the analysis.</p>
				<p>In earlier work on speaker discrimination by speech sound (Andics et al., 2007), responses from same-speaker trials were analyzed separately from responses to different-speaker trials. We chose to analyze all data together, but added trial type (same-speaker, different-speaker) as a factor into the design. The other fixed factors in the design were focus (+, &#x2212;), vowel nucleus ([a&#x02D0;], [u]), and word context ([v_t], [&#x0266;_k]), with full interactions. All factors were coded using deviation coding. Random intercepts by participants and by trials (that is, speaker1&#x2013;speaker2 combinations) were included in the design, as were analogous random slopes for all factor combinations included as fixed effects. A diagonal random-effects covariance matrix was assumed. The dependent variable was the correctness of the response given by the listener, coded as 0 for incorrect and 1 for correct. The data were analyzed using a mixed-effects logistic-regression tree (Fokkema et al., 2018; see Tagliamonte &#x0026; Baayen, 2012 for an accessible introduction to tree-based models in linguistics). Function buildmertree from R package buildmer (Voeten, 2020) was used to find the maximal random-effects structure that still converged non-singularly, based on the random effects&#x2019; contributions to the AIC (Akaike, 1971) of the model. The results of the perception experiment are presented in section 3.2.</p>
			</sec>
		</sec>
		<sec sec-type="heading-1">
			<label>3</label>
			<title>Results</title>
			<sec sec-type="heading-2">
				<label>3.1</label>
				<title>Stimulus acoustics: the effect of focus condition</title>
				<p>Before a perceptual effect of [&#x00B1;focus] on speaker discrimination was evaluated, the effect of [&#x00B1;focus] realizations on stimulus acoustics was assessed. Per acoustic measure, the model output on effects of Focus, Vowel and Word Context is given in Table&#x00A0;1. Significant terms are printed in bold. Three out of four acoustic measurements showed a main effect of or interaction with Focus; these mainly reflected a higher F0 and longer stimulus durations in focused realizations. The effect was not equally strong in all stimulus conditions. As for F1, there was a tendency for more mouth opening in [a&#x02D0;] (given the marginal main effect main effect of Focus, <italic>t</italic> = 2.2), but in [u] this effect was countered under focus, presumably because of increased rounding in its pronunciation. Moreover, stimulus acoustics were influenced by vowel quality, as may be expected by intrinsic differences between the vowels, and to a lesser extent by the word context and interactions between the linguistic factors.</p>
				<p>Table&#x00A0;2 shows, in a descriptive manner, the within-speaker variance by condition and by acoustic measure as well as the speaker-specificity index (SSI, the ratio of between-to-within-speaker variances). The F0 has higher within-speaker variation in focused than unfocused words, but also a higher SSI meaning that between-speaker variances also increase with focus. F1 shows less within-speaker variance with focus, whereas SSI tends to increase. F2 shows an increase in within-speaker variance with focus in the [v_t] condition, but a decrease in [&#x0266;_k], whereas SSI shows behavior in the opposite direction. For [a&#x02D0;] duration, but not for [u], SSI is higher when focused.</p>
				<table-wrap>
					<caption>
						<title>Table&#x00A0;1</title>
						<p>Modelling results for the acoustic parameters log-duration, and F0, F1 and F2 (all three in Bark, ***: <italic>p</italic> &#x003C; .001, **: <italic>p</italic> &#x003C; .01, *: <italic>p</italic> =.01). Significant terms in bold</p>
					</caption>
					<table frame="void" cellpadding="5">
						<thead>
							<tr>
								<th>
									<p></p>
								</th>
								<th colspan="2">
									<p>F0 [Bark]</p>
								</th>
								<th colspan="2">
									<p>Log(duration)</p>
								</th>
							</tr>
							<tr>
								<th>
									<p></p>
								</th>
								<th>
									<p>Estimate</p>
								</th>
								<th>
									<p><italic>t</italic>-value</p>
								</th>
								<th>
									<p>Estimate</p>
								</th>
								<th>
									<p><italic>t</italic>-value</p>
								</th>
							</tr>
						</thead>
						<tbody>
							<tr>
								<td>
									<p>Intercept</p>
								</td>
								<td>
									<p><bold>0</bold><bold>.699</bold></p>
								</td>
								<td>
									<p>18.0***</p>
								</td>
								<td>
									<p><bold>&#x2212;0</bold><bold>.626</bold></p>
								</td>
								<td>
									<p>&#x2212;28.0***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]</p>
								</td>
								<td>
									<p><bold>0</bold><bold>.448</bold></p>
								</td>
								<td>
									<p>5.8***</p>
								</td>
								<td>
									<p><bold>0</bold><bold>.079</bold></p>
								</td>
								<td>
									<p>4.0***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Vowel_[u]</p>
								</td>
								<td>
									<p><bold>0</bold><bold>.141</bold></p>
								</td>
								<td>
									<p>4.8***</p>
								</td>
								<td>
									<p>&#x2212;0.009</p>
								</td>
								<td>
									<p>&#x2212;0.7</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>WordContext_[v_t]</p>
								</td>
								<td>
									<p><bold>0</bold><bold>.145</bold></p>
								</td>
								<td>
									<p>5.0***</p>
								</td>
								<td>
									<p><bold>0</bold><bold>.093</bold></p>
								</td>
								<td>
									<p>5.1***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]: Vowel_[u]</p>
								</td>
								<td>
									<p>0.011</p>
								</td>
								<td>
									<p>0.3</p>
								</td>
								<td>
									<p><bold>&#x2212;0</bold><bold>.063</bold></p>
								</td>
								<td>
									<p>&#x2212;3.5***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]: WordContext_[v_t]</p>
								</td>
								<td>
									<p><bold>&#x2212;0</bold><bold>.211</bold></p>
								</td>
								<td>
									<p>&#x2212;5.2***</p>
								</td>
								<td>
									<p>&#x2212;0.019</p>
								</td>
								<td>
									<p>&#x2212;1.1</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Vowel_[u]: WordContext_[v_t]</p>
								</td>
								<td>
									<p>&#x2212;0.058</p>
								</td>
								<td>
									<p>&#x2212;1.4</p>
								</td>
								<td>
									<p><bold>&#x2212;0</bold><bold>.116</bold></p>
								</td>
								<td>
									<p>&#x2212;6.5***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]: Vowel_[u]: WordContext_[v_t]</p>
								</td>
								<td>
									<p>0.102</p>
								</td>
								<td>
									<p>1.8</p>
								</td>
								<td>
									<p><bold>0</bold><bold>.097</bold></p>
								</td>
								<td>
									<p>3.9***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td colspan="2">
									<p><bold>F1 [Bark]</bold></p>
								</td>
								<td colspan="2">
									<p><bold>F2 [Bark]</bold></p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p><bold>Estimate</bold></p>
								</td>
								<td>
									<p><bold><italic>t</italic>-value</bold></p>
								</td>
								<td>
									<p><bold>Estimate</bold></p>
								</td>
								<td>
									<p><bold><italic>t</italic>-value</bold></p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Intercept</p>
								</td>
								<td>
									<p><bold>6</bold><bold>.942</bold></p>
								</td>
								<td>
									<p>52.6***</p>
								</td>
								<td>
									<p><bold>10</bold><bold>.391</bold></p>
								</td>
								<td>
									<p>265.2***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]</p>
								</td>
								<td>
									<p>0.401</p>
								</td>
								<td>
									<p>2.2</p>
								</td>
								<td>
									<p>0.107</p>
								</td>
								<td>
									<p>0.5</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Vowel_[u]</p>
								</td>
								<td>
									<p><bold>&#x2212;3</bold><bold>.803</bold></p>
								</td>
								<td>
									<p>&#x2212;15.7***</p>
								</td>
								<td>
									<p><bold>&#x2212;2</bold><bold>.213</bold></p>
								</td>
								<td>
									<p>&#x2212;5.4***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>WordContext_[v_t]</p>
								</td>
								<td>
									<p>&#x2212;0.277</p>
								</td>
								<td>
									<p>&#x2212;1.8</p>
								</td>
								<td>
									<p>&#x2212;0.380</p>
								</td>
								<td>
									<p>&#x2212;1.7</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]: Vowel_[u]</p>
								</td>
								<td>
									<p><bold>&#x2212;0</bold><bold>.600</bold></p>
								</td>
								<td>
									<p>&#x2212;2.8**</p>
								</td>
								<td>
									<p>&#x2212;0.640</p>
								</td>
								<td>
									<p>&#x2212;2.1</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]: WordContext_[v_t]</p>
								</td>
								<td>
									<p>&#x2212;0.143</p>
								</td>
								<td>
									<p>&#x2212;0.7</p>
								</td>
								<td>
									<p>0.013</p>
								</td>
								<td>
									<p>0.04</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Vowel_[u]: WordContext_[v_t]</p>
								</td>
								<td>
									<p>0.053</p>
								</td>
								<td>
									<p>0.3</p>
								</td>
								<td>
									<p><bold>1</bold><bold>.171</bold></p>
								</td>
								<td>
									<p>3.8***</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>Focus_[+]: Vowel_[u]: WordContext_[v_t]</p>
								</td>
								<td>
									<p>0.611</p>
								</td>
								<td>
									<p>2.0</p>
								</td>
								<td>
									<p>0.354</p>
								</td>
								<td>
									<p>0.8</p>
								</td>
							</tr>
						</tbody>
					</table>
				</table-wrap>
				<table-wrap>
					<caption>
						<title>Table&#x00A0;2</title>
						<p>Mean within-speaker variance and speaker-specificity index (SSI) per acoustic parameter. Measures are given by vowel, focus condition and word context</p>
					</caption>
					<table frame="void" cellpadding="5">
						<thead>
							<tr>
								<th>
									<p></p>
								</th>
								<th>
									<p></p>
								</th>
								<th>
									<p>[&#x2212;focus]</p>
								</th>
								<th>
									<p></p>
								</th>
								<th>
									<p>[+focus]</p>
								</th>
							</tr>
							<tr>
								<th>
									<p></p>
								</th>
								<th>
									<p>Parameter</p>
								</th>
								<th>
									<p>[&#x0266;_k]</p>
								</th>
								<th>
									<p>[v_t]</p>
								</th>
								<th>
									<p>[&#x0266;_k]</p>
								</th>
								<th>
									<p>[v_t]</p>
								</th>
							</tr>
						</thead>
						<tbody>
							<tr>
								<td>
									<p>[a&#x02D0;]</p>
								</td>
								<td>
									<p>F0_within</p>
								</td>
								<td>
									<p>161</p>
								</td>
								<td>
									<p>163</p>
								</td>
								<td>
									<p>223</p>
								</td>
								<td>
									<p>136</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F0_SSI</p>
								</td>
								<td>
									<p>1.03</p>
								</td>
								<td>
									<p>1.39</p>
								</td>
								<td>
									<p>3.09</p>
								</td>
								<td>
									<p>3.65</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F1_within</p>
								</td>
								<td>
									<p>1,940</p>
								</td>
								<td>
									<p>1,419</p>
								</td>
								<td>
									<p>1,224</p>
								</td>
								<td>
									<p>902</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F1_SSI</p>
								</td>
								<td>
									<p>2.12</p>
								</td>
								<td>
									<p>1.39</p>
								</td>
								<td>
									<p>2.42</p>
								</td>
								<td>
									<p>2.80</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F2_within</p>
								</td>
								<td>
									<p>6,363</p>
								</td>
								<td>
									<p>1,774</p>
								</td>
								<td>
									<p>2,849</p>
								</td>
								<td>
									<p>1,989</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F2_SSI</p>
								</td>
								<td>
									<p>1.41</p>
								</td>
								<td>
									<p>3.34</p>
								</td>
								<td>
									<p>2.02</p>
								</td>
								<td>
									<p>2.45</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>Duration_within</p>
								</td>
								<td>
									<p>0.001</p>
								</td>
								<td>
									<p>0.001</p>
								</td>
								<td>
									<p>0.001</p>
								</td>
								<td>
									<p>0.001</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>Duration_SSI</p>
								</td>
								<td>
									<p>2.33</p>
								</td>
								<td>
									<p>2.33</p>
								</td>
								<td>
									<p>4.60</p>
								</td>
								<td>
									<p>3.99</p>
								</td>
							</tr>
							<tr>
								<td>
									<p>[u]</p>
								</td>
								<td>
									<p>F0_within</p>
								</td>
								<td>
									<p>76</p>
								</td>
								<td>
									<p>73</p>
								</td>
								<td>
									<p>136</p>
								</td>
								<td>
									<p>152</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F0_SSI</p>
								</td>
								<td>
									<p>2.22</p>
								</td>
								<td>
									<p>2.60</p>
								</td>
								<td>
									<p>8.05</p>
								</td>
								<td>
									<p>3.62</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F1_within</p>
								</td>
								<td>
									<p>1,753</p>
								</td>
								<td>
									<p>905</p>
								</td>
								<td>
									<p>836</p>
								</td>
								<td>
									<p>712</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F1_SSI</p>
								</td>
								<td>
									<p>1.16</p>
								</td>
								<td>
									<p>2.79</p>
								</td>
								<td>
									<p>1.86</p>
								</td>
								<td>
									<p>2.39</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F2_within</p>
								</td>
								<td>
									<p>9,138</p>
								</td>
								<td>
									<p>4,509</p>
								</td>
								<td>
									<p>2,276</p>
								</td>
								<td>
									<p>7,627</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>F2_SSI</p>
								</td>
								<td>
									<p>0.95</p>
								</td>
								<td>
									<p>3.01</p>
								</td>
								<td>
									<p>1.70</p>
								</td>
								<td>
									<p>1.41</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>Duration_within</p>
								</td>
								<td>
									<p>0.005</p>
								</td>
								<td>
									<p>0.001</p>
								</td>
								<td>
									<p>0.004</p>
								</td>
								<td>
									<p>0.002</p>
								</td>
							</tr>
							<tr>
								<td>
									<p></p>
								</td>
								<td>
									<p>Duration_SSI</p>
								</td>
								<td>
									<p>1.53</p>
								</td>
								<td>
									<p>2.1</p>
								</td>
								<td>
									<p>1.55</p>
								</td>
								<td>
									<p>1.75</p>
								</td>
							</tr>
						</tbody>
					</table>
				</table-wrap>
				<p>Together, these results show that stimulus acoustics as well as the SSI vary by focus condition, supporting the hypothesis of differential acoustic speaker information by factor combination.</p>
			</sec>
			<sec sec-type="heading-2">
				<label>3.2</label>
				<title>Perception results: effects of focus condition and vowel</title>
				<p>The main research questions on speaker discrimination as a function of focus condition and vowel quality were evaluated using a same-different perception task. The results of the statistical analysis are shown in Figure&#x00A0;1, and reflect that listeners were generally quite successful in speaker discrimination, but also that performance varied by factor combination. The first, and hence most important, split in the tree model is on vowel, [a&#x02D0;] versus [u]. On [a&#x02D0;] trials listeners gave 87.5&#x202F;% correct responses, irrespective of focus, word context or trial type. The bar charts show that performance on [u] conditions was somewhat lower.</p>
				<fig orientation="portrait" position="float">
					<caption>
						<title>Figure&#x00A0;1</title>
						<p>The fitted mixed-effects logistic-regression tree, modelling participants&#x2019; correct responses as a function of Vowel ([a&#x02D0;]-[u]), Focus ([+/-]), Word Context ([&#x0266;_k], [v_t]), and Trial Type (DS = different speaker, SS = same speaker)</p>
					</caption>
					<graphic xlink:href="DuJAL9420_i0001.png"/>
				</fig>
				<p>Within [u] trials, the first split was on the predictor Focus and performance further depended on the combination of word context and trial type. For [&#x2212;focus] words containing [u], Word Context was the next split, followed by Trial Type on both contexts. Different-speaker trials (75.1&#x202F;%) were better than same-speaker trials (68.2&#x202F;%) containing non-focused tokens of [&#x0266;uk]. For non-focused tokens of [vut] same-speaker trials (84.0&#x202F;%) were better than different-speaker trials (77.3&#x202F;%).</p>
				<p>For [+focus] words containing [u], the first split was on Trial Type, followed by a split on Word Context for different-speaker trials only. On same-speaker trials, listeners got 88.5&#x202F;% correct responses. On different-speaker trials, [vut] tokens received more correct responses (80.7&#x202F;%) than [&#x0266;uk] tokens (73.9&#x202F;%).</p>
				<p>The model also contained a number of random effects of participant and of trial, including random slopes. To further investigate the random effects and thus look for potential confounds, a cluster analysis was applied, following the method in Voeten (2020). No interpretable<xref ref-type="fn" rid="FN000001"><sup>1</sup></xref> structure was found in the by-participants or by-items random-effects structure, which suggests that the experiment was free from confounds due to these random effects.</p>
			</sec>
		</sec>
		<sec sec-type="heading-1">
			<label>4</label>
			<title>Discussion and conclusion</title>
			<p>This study investigated if speaker discrimination performance varies with the presence of focus on target words and if the difference in speaker-specific information between the Dutch corner vowels [a&#x02D0;] and [u] found in acoustic analyses (Van den Heuvel, 1996) is also observed in perception. On average, listeners reached over 80&#x202F;% correct responses for both same-speaker and different-speaker trials. The regression tree further showed that different factor combinations, reflecting different stimulus acoustics, affected the listeners&#x2019; ability to discriminate speakers.</p>
			<p>In line with expectations, listeners gave more correct responses to stimulus pairs containing [a&#x02D0;] words than [u] words. These results complement the perceptual results obtained by Andics et al. (2007) on Dutch vowels [&#x03B5;] and [&#x0254;], and follow the literature on acoustic speaker information in [a&#x02D0;] versus [u] in Dutch (Van den Heuvel, 1996), and also in other languages (Fejlov&#x00E1; et al., 2013; Schindler &#x0026; Draxler, 2013). Perception of tokens containing the vowel [a&#x02D0;] was not influenced further by linguistic context. For [u], the situation was different: the first split in the model was made by Focus, where focused realizations reached higher discrimination performance than unfocused ones. Best performance was observed for [+focus] same-speaker trials, where the phonetic-acoustic differences at the surface are apparently small enough for good-quality voice matching. This result shows that the phonetic-acoustic content resulting from the linguistic position in which a word is pronounced may influence the speaker-dependent information available to listeners. Yet, the effect obtained here is restricted to the less informative vowel, [u].</p>
			<p>In the literature it has been reported that listeners tend to underestimate within-speaker variation when listening to unfamiliar speakers (Lavan et al., 2019), and are thus more likely to perceive same-speaker samples as coming from different speakers. This would especially lower listeners&#x2019; accuracy in conditions where within-speaker variance is high. The finding that listeners performed better using focused than unfocused [u], however, suggests that the mere amount of within-speaker variance does not fully explain the discrimination results; as Table&#x00A0;2 shows, within-speaker variance of F0, an important parameter in speaker discrimination (Baumann &#x0026; Belin, 2010), tended to be larger in focused than unfocused words. At the same time, however, the SSI&#x202F;s of F0 showed that there was also a tendency for between-speaker variance to increase under focus, and also to increase more than within-speaker variance. This suggests that listeners use both within-speaker and between-speaker variance in perception.</p>
			<p>Additional evidence that within-speaker variance alone may not explain the speaker discrimination results comes from the comparison of listener performance on [a&#x02D0;] versus [u]. For F1, within-speaker variances are larger for [a&#x02D0;], whereas for F2, within-speaker variances are larger for [u]. SSI, however, is in most cases larger in [a&#x02D0;] than [u], again suggesting a contribution for both sources of variation in perception. The acoustic analysis in this study was limited to one vocal source and two vocal filter parameters, and these are only a small subset of the parameters that have been included in the recent literature (Lee et al., 2019). That work furthermore showed an important contribution of measures capturing phonetic change within stimuli, which was not included here. More detailed analyses of which information listeners use to perform speaker discrimination tasks are left for future research.</p>
			<p>The two splits on Word Context, within [u], reflect that speaker discrimination was better on [vut] than [&#x0266;uk] contexts, even though word contexts had been selected to contain similar amounts of acoustic speaker information. As plosives have been reported to contain the least speaker-dependent information (Van den Heuvel, 1996), the word-context effect must be attributed mainly to the differential information contained by the onsets. Schindler and Draxler (2013) demonstrated that there was slightly more speaker-dependent information in the spectra of [v] than of [h]. In addition, in the current investigation [v] words in most cases were longer than [&#x0266;] words (see Table&#x00A0;1), where additional duration is generally beneficial in perceptual tasks (Bricker &#x0026; Pruzansky, 1966; Orchard &#x0026; Yarmey, 1995; Cook &#x0026; Wilding, 1997). Finally, the two fricatives are expected to differ in coarticulation with the vowel, with [v] presumably showing larger between-speaker differences in coarticulation than [&#x0266;]. This may also contribute to explaining the word context difference (Bricker &#x0026; Pruzansky, 1966; Lee et al., 2019).</p>
			<p>The current investigation was limited in its scope, studying the effect of focus using two vowels, in two word contexts each. The higher-order interactions in the results showed that listeners are sensitive to subtle differences conditioned by linguistic structure when processing voice information. However, in the current experiment very short utterances were presented, produced by a relatively homogeneous group of male speakers. In natural communicative settings, more variation between (non-seen) speakers is likely to occur, and is expected to improve speaker discrimination. Those circumstances potentially reduce the effect of relatively subtle linguistic cues; such cues here affected speaker discrimination only in the more-challenging words containing [u], but not those containing [a&#x02D0;]. Moreover, if longer utterances had been used, as would be found in natural interaction, listeners would receive additional speaker information from the utterance. However, if within-speaker variability increases over longer phrases, speaker discriminability may also be compromised.</p>
			<p>To conclude, an effect of focus on speaker-dependent information contained by a word was present, but limited. Moreover, additional evidence was found for the claim that different speech sounds differ in speaker-dependent information ([a&#x02D0;] versus [u], and possibly [v] versus [&#x0266;]). At the same time, under more real-world conditions than the current discrimination task, the detection of speaker changes in speech is unlikely to be affected by focus. When needed, however, listeners seem skilled at exploiting the little information that is available.</p>
		</sec>
		<sec sec-type="heading-1">
			<title>Acknowledgements</title>
			<p>This research was supported by a VIDI grant (276-75-010) from the Dutch Research Council.</p>
		</sec>
		<sec sec-type="heading-1">
			<title>Notes</title>
		</sec>
		<sec sec-type="heading-1">
			<title>Appendix A: Statement-question pairs used to elicit [&#x00B1;focus] versions of the target words</title>
			<p>Wat stoot hij aan de tafel? Hij stoot zijn <bold>voet</bold> aan de tafel</p>
			<p>&#x2018;<italic>How did he hit the table? He bumped his foot into the table</italic>&#x2019;</p>
			<p>Waar stoot hij zijn voet aan? Hij stoot zijn voet aan de <bold>tafel</bold></p>
			<p>&#x2018;<italic>How did he hit his foot? He bumped his foot into the table</italic>&#x2019;</p>
			<p>Wat zet hij in de wasbak? Hij zet zijn <bold>vaat</bold> in de wasbak</p>
			<p>&#x2018;<italic>What does he place in the sink? He puts his dishes in the sink</italic>&#x2019;</p>
			<p>Waar zet hij zijn vaat? Hij zet zijn vaat in de <bold>wasbak</bold></p>
			<p>&#x2018;<italic>Where does he put his dishes? He puts his dishes in the sink</italic>&#x2019;</p>
			<p>Waar staat ze te wachten? Ze staat op de <bold>hoek</bold> te wachten</p>
			<p>&#x2018;<italic>Where is she waiting? She&#x2019;s waiting at the corner</italic>&#x2019;</p>
			<p>Wat staat ze te doen op de hoek? Ze staat op de hoek te <bold>wachten</bold></p>
			<p>&#x2018;<italic>What is she doing at the corner? She&#x2019;s waiting at the corner</italic>&#x2019;</p>
			<p>Waar is de vis aan geslagen? De vis is aan de <bold>haak</bold> geslagen</p>
			<p>&#x2018;<italic>With what was the fish caught? The fish was caught on the hook</italic>&#x2019;</p>
			<p>Wat is er aan de haak geslagen? De <bold>vis</bold> is aan de haak geslagen</p>
			<p>&#x2018;<italic>What was caught on the hook? The fish was caught on the hook</italic>&#x2019;</p>
		</sec>
		<sec sec-type="heading-1">
			<title>Appendix B: Visualization of the two significant clusters for the by-trial random slope for the interaction Focus by Vowel</title>
			<fig orientation="portrait" position="anchor">
				<graphic xlink:href="DuJAL9420_i0002.png"/>
			</fig>
		</sec>
	</body>
	<back>
		<fn-group>
			<fn id="FN000001"><label>1</label><p>Two significant clusters were identified for the by-trial random slope for the interaction Focus by Vowel, but the resulting two clusters (see Appendix B) seem to be spurious. We note that Appendix B&#x2019;s Cluster 2 seems to be relatively tightly concentrated around zero, while Cluster 1 seems to stay farther away, but we do not have any meaningful interpretation for these differences.</p></fn>
		</fn-group>
		<ref-list>
			<title>References</title>
			<ref id="R000001">
				<mixed-citation>Akaike, H. (1971). Information theory and an extension of the maximum likelihood principle. In B.N. Petrov, &#x0026; F. Csáki (Eds.), <italic>2nd International Symposium on Information Theory, Tsahkadsor, Armenia, USSR</italic>, September 2–8, 1971 (pp. 267–281). Budapest, Akadémiai Kiadó.</mixed-citation>
			</ref>
			<ref id="R000002">
				<mixed-citation>Amino, K., &#x0026; Arai, T. (2007). Contribution of consonants and vowels to the perception of speaker identity. In <italic>Japan-China Joint Conference on Acoustics</italic>. Sendai, Japan.</mixed-citation>
			</ref>
			<ref id="R000003">
				<mixed-citation>Andics, A., McQueen, J.M., &#x0026; Van Turennout, M. (2007). Phonetic content influences voice discriminability. In J. Trouvain, &#x0026; W.J. Barry (Eds.), <italic>Proceedings of the 16th International Congress of Phonetic Sciences</italic> (pp. 1829–1832). Dudweiler: Pirrot. [<ext-link xlink:href="https://pure.mpg.de/pubman/faces/ViewItemOverviewPage.jsp?itemId=item_57725">https://pure.mpg.de/pubman/faces/ViewItemOverviewPage.jsp?itemId=item_57725</ext-link>]</mixed-citation>
			</ref>
			<ref id="R000004">
				<mixed-citation>Baumann, O., &#x0026; Belin, P. (2010). Perceptual scaling of voice identity: Common dimensions for different vowels and speakers. <italic>Psychological Research</italic> PRPF, 74(1), 110. [<ext-link xlink:href="https://link.springer.com/article/10.1007/s00426-008-0185-z">https://link.springer.com/article/10.1007/s00426-008-0185-z</ext-link>]</mixed-citation>
			</ref>
			<ref id="R000005">
				<mixed-citation>Bell, A., Brenier, J.M., Gregory, M., Girand, C., &#x0026; Jurafsky, D. (2009). Predictability effects on durations of content and function words in conversational English. <italic>Journal of Memory and Language</italic>, 60(1), 92–111. <ext-link xlink:href="https://doi.org/10.1016/j.jml.2008.06.003">https://doi.org/10.1016/j.jml.2008.06.003</ext-link></mixed-citation>
			</ref>
			<ref id="R000006">
				<mixed-citation>Boersma, P., &#x0026; Weenink, D. (2018). <italic>Praat. Doing phonetics by computer</italic> (Version 6.0.42 ) [Computer program].</mixed-citation>
			</ref>
			<ref id="R000007">
				<mixed-citation>Bricker, P.D., &#x0026; Pruzansky, S. (1966). Effects of stimulus content and duration on talker identification. <italic>The Journal of the Acoustical Society of America</italic>, 40, 1441–1449. <ext-link xlink:href="https://doi.org/10.1121/1.1910246">https://doi.org/10.1121/1.1910246</ext-link></mixed-citation>
			</ref>
			<ref id="R000008">
				<mixed-citation>Chafe, W.L. (1976). Givenness, contrastiveness, definiteness, subjects, topics and point of view. In C.N. Li (Ed.) <italic>Subject and topic</italic> (pp. 27–55). Academic Press.</mixed-citation>
			</ref>
			<ref id="R000009">
				<mixed-citation>Chen, A. (2009). The phonetics of sentence-initial topic and focus in adult and child Dutch. In M. Vigário, S. Frota, &#x0026; M. João Freitas (Eds.), <italic>Phonetics and phonology: interactions and interrelations</italic> (pp. 91–106). John Benjamins Publishing Company.</mixed-citation>
			</ref>
			<ref id="R000010">
				<mixed-citation>Cook, S., &#x0026; Wilding, J. (1997). Earwitness testimony: Never mind the variety, hear the length. <italic>Applied Cognitive Psychology</italic>, 11(2), 95–111. <ext-link xlink:href="https://doi.org/10.1002/(SICI)1099&#x2013;0720(199704)11:2&#x003C;95::AID-ACP429&#x003E;3.0.CO;2-O">https://doi.org/10.1002/(SICI)1099–0720(199704)11:2&#x003C;95::AID-ACP429&#x003E;3.0.CO;2-O</ext-link></mixed-citation>
			</ref>
			<ref id="R000011">
				<mixed-citation>Drozdova, P., Van Hout, R., &#x0026; Scharenborg, O. (2017). L2 voice recognition: The role of speaker-, listener-, and stimulus-related factors. <italic>The Journal of the Acoustical Society of America</italic>, 142(5), 3058–3068. <ext-link xlink:href="https://doi.org/10.1121/1.5010169">https://doi.org/10.1121/1.5010169</ext-link></mixed-citation>
			</ref>
			<ref id="R000012">
				<mixed-citation>Eefting, W. (1991). The effect of “information value” and “accentuation” on the duration of Dutch words, syllables, and segments. <italic>The Journal of the Acoustical Society of America</italic>, 89(1), 412–424.</mixed-citation>
			</ref>
			<ref id="R000013">
				<mixed-citation>Fant, G. (1960). <italic>Acoustic theory of speech production</italic>. Mouton and Co.</mixed-citation>
			</ref>
			<ref id="R000014">
				<mixed-citation>Fejlová, D., Lukeš, D., &#x0026; Skarnitzl, R. (2013). Formant contours in Czech vowels: Speaker-discriminating potential. <italic>Proceedings of Interspeech 2013</italic>, 25–29 August 2013, Lyon, France (pp. 3182–3186).</mixed-citation>
			</ref>
			<ref id="R000015">
				<mixed-citation>Fokkema, M., Smits, N., Zeileis, A., Hothorn, T., &#x0026; Kelderman, H. (2018). Detecting treatment-subgroup interactions in clustered data with generalized linear mixed-effects model trees. <italic>Behavior Research Methods</italic>, 50, 2016–2034. [<ext-link xlink:href="https://link.springer.com/article/10.3758/s13428-017-0971-x">https://link.springer.com/article/10.3758/s13428-017-0971-x</ext-link>]</mixed-citation>
			</ref>
			<ref id="R000016">
				<mixed-citation>Hanssen, J., Peters, J., &#x0026; Gussenhoven, C. (2008). Prosodic effects of focus in Dutch declaratives. <italic>Proceedings of the 4th international conference on Speech Prosody</italic>. Campinas, Brazil, pp. 609–612.</mixed-citation>
			</ref>
			<ref id="R000017">
				<mixed-citation>He, L., &#x0026; Dellwo, V. (2017). Between-speaker variability in temporal organizations of intensity contours. <italic>The Journal of the Acoustical Society of America</italic>, 141(5), EL488–EL494. <ext-link xlink:href="https://doi.org/10.1121/1.4983398">https://doi.org/10.1121/1.4983398</ext-link></mixed-citation>
			</ref>
			<ref id="R000018">
				<mixed-citation>He, L., Zhang, Y., &#x0026; Dellwo, V. (2019). Between-speaker variability and temporal organization of the first formant. <italic>The Journal of the Acoustical Society of America</italic>, 145(3), EL209–EL214. <ext-link xlink:href="https://doi.org/10.1121/1.5093450">https://doi.org/10.1121/1.5093450</ext-link></mixed-citation>
			</ref>
			<ref id="R000019">
				<mixed-citation>Heeren W.F.L. (2020). The effect of word class on speaker-dependent information in the Standard Dutch vowel /aː/. <italic>The Journal of the Acoustical Society of America</italic>, 148(4), 2028–2039. <ext-link xlink:href="https://doi.org/10.1121/10.0002173">https://doi.org/10.1121/10.0002173</ext-link></mixed-citation>
			</ref>
			<ref id="R000020">
				<mixed-citation>Kavanagh, C. (2012). <italic>New consonantal acoustic parameters for forensic speaker comparison</italic> Doctoral dissertation. University of York. [<ext-link xlink:href="https://etheses.whiterose.ac.uk/3980/">https://etheses.whiterose.ac.uk/3980/</ext-link></mixed-citation>
			</ref>
			<ref id="R000021">
				<mixed-citation>Keuleers, E., Brysbaert, M., &#x0026; New, B. (2010). SUBTLEX-NL: A new frequency measure for Dutch words based on film subtitles. <italic>Behavior Research Methods</italic>, 42(3), 643–650. <ext-link xlink:href="https://doi.org/10.3758/BRM.42.3.643">https://doi.org/10.3758/BRM.42.3.643</ext-link></mixed-citation>
			</ref>
			<ref id="R000022">
				<mixed-citation>Krifka, M. (2007). Basic notions of information structure. In C. Féry, G. Fanselow, &#x0026; M. Krifka (Eds.), <italic>The notions of information structure</italic> (pp. 13–55). Universitätsverlag Potsdam.</mixed-citation>
			</ref>
			<ref id="R000023">
				<mixed-citation>Lavan, N., Burston, L.F.K., &#x0026; Garrido, L. (2019). How many voices did you hear? Natural variability disrupts identity perception from unfamiliar voices. <italic>British Journal of Psychology</italic>, 110, 576–593. <ext-link xlink:href="https://doi.org/10.1111/bjop.12348">https://doi.org/10.1111/bjop.12348</ext-link></mixed-citation>
			</ref>
			<ref id="R000024">
				<mixed-citation>Lee, Y., Keating, P., &#x0026; Kreiman, J. (2019). Acoustic voice variation within and between speakers. <italic>The Journal of the Acoustical Society of America</italic>, 146(3), 1568–1579. <ext-link xlink:href="https://doi.org/10.1121/1.5125134">https://doi.org/10.1121/1.5125134</ext-link></mixed-citation>
			</ref>
			<ref id="R000025">
				<mixed-citation>McDougall, K. (2006). Dynamic features of speech and the characterization of speakers: Towards a new approach using formant frequencies. <italic>International Journal of Speech, Language and the Law</italic>, 13(1), 89–126.</mixed-citation>
			</ref>
			<ref id="R000026">
				<mixed-citation>Morrison, G.S. (2009). Likelihood-ratio forensic voice comparison using parametric representations of the formant trajectories of diphthongs. <italic>The Journal of the Acoustical Society of America</italic>, 125(4), 2387–2397. <ext-link xlink:href="https://doi.org/10.1121/1.3081384">https://doi.org/10.1121/1.3081384</ext-link></mixed-citation>
			</ref>
			<ref id="R000027">
				<mixed-citation>Narayan, C.R., Mak, L., &#x0026; Bialystok, E. (2017). Words get in the way: Linguistic effects on talker discrimination. <italic>Cognitive Science</italic>, 41(5), 1361–1376. <ext-link xlink:href="https://doi.org/10.1111/cogs.12396">https://doi.org/10.1111/cogs.12396</ext-link></mixed-citation>
			</ref>
			<ref id="R000028">
				<mixed-citation>Orchard, T., &#x0026; Yarmey, A.D. (1995). The effects of whispers, voice-sample duration, and voice distinctiveness on criminal speaker identification, <italic>Applied Cognitive Psychology</italic>, 9(3), 249–260. <ext-link xlink:href="https://doi.org/10.1002/acp.2350090306">https://doi.org/10.1002/acp.2350090306</ext-link></mixed-citation>
			</ref>
			<ref id="R000029">
				<mixed-citation>Psychology Software Tools. (2012). <italic>E-Prime (Version 2.0)</italic>. <ext-link xlink:href="https://www.pstnet.com">https://www.pstnet.com</ext-link></mixed-citation>
			</ref>
			<ref id="R000030">
				<mixed-citation>Schindler, C., &#x0026; Draxler, C. (2013) Using spectral moments as a speaker specific feature in nasals and fricatives. <italic>Proceedings of Interspeech</italic> (pp. 2793–2796), Lyon, France, 25–29 August 2013.</mixed-citation>
			</ref>
			<ref id="R000031">
				<mixed-citation>Sluijter, A.M.C., &#x0026; Van Heuven, V.J. (1996). Spectral balance as an acoustic correlate of linguistic stress. <italic>The Journal of the Acoustical Society of America</italic>, 100, 2471–2485. <ext-link xlink:href="https://doi.org/10.1121/1.417955">https://doi.org/10.1121/1.417955</ext-link></mixed-citation>
			</ref>
			<ref id="R000032">
				<mixed-citation>Smorenburg, B.J.L., &#x0026; Heeren, W.F.L. (2020). The distribution of speaker information in Dutch fricatives /s/ and /x/ from telephone dialogues. <italic>Journal of the Acoustical Society of America</italic>, 147(2), 949–960. <ext-link xlink:href="https://doi.org/10.1121/10.0000674">https://doi.org/10.1121/10.0000674</ext-link></mixed-citation>
			</ref>
			<ref id="R000033">
				<mixed-citation>Smorenburg, B.J.L., &#x0026; Heeren W.F.L. (2021). Acoustic and speaker variation in Dutch /n/ and /m/ as a function of phonetic context and syllabic position. <italic>The Journal of the Acoustical Society of America</italic>, 150(2), 979–989. <ext-link xlink:href="https://doi.org/10.1121/10.0005845">https://doi.org/10.1121/10.0005845</ext-link></mixed-citation>
			</ref>
			<ref id="R000034">
				<mixed-citation>Stevenage, S.V. (2018). Drawing a distinction between familiar and unfamiliar voice processing: A review of neuropsychological, clinical and empirical findings. <italic>Neuropsychologia</italic>, 116, 162–178. <ext-link xlink:href="https://doi.org/10.1016/j.neuropsychologia.2017.07.005">https://doi.org/10.1016/j.neuropsychologia.2017.07.005</ext-link></mixed-citation>
			</ref>
			<ref id="R000035">
				<mixed-citation>Tagliamonte, S.A., &#x0026; Baayen, R.H. (2012). Models, forests, and trees of York English: Was/were variation as a case study for statistical practice. <italic>Language Variation and Change</italic>, 24, 135–178. [<ext-link xlink:href="https://clarinoai.informatik.uni-leipzig.de/fedora/objects/oai:mrr/datastreams/info/content">https://clarinoai.informatik.uni-leipzig.de/fedora/objects/oai:mrr/datastreams/info/content</ext-link>]</mixed-citation>
			</ref>
			<ref id="R000036">
				<mixed-citation>Van Bergem, D.R. (1993). Acoustic vowel reduction as a function of sentence accent, word stress, and word class. <italic>Speech Communication</italic>, 12(1), 1–23. <ext-link xlink:href="https://doi.org/10.1016/0167-6393(93)90015-D">https://doi.org/10.1016/0167-6393(93)90015-D</ext-link></mixed-citation>
			</ref>
			<ref id="R000037">
				<mixed-citation>Van Bergem, D.R. (1995). Acoustic and lexical vowel reduction. PhD dissertation, University of Amsterdam. [<ext-link xlink:href="https://dare.uva.nl/search?identifier=6ba47af3-8bf4-4b46-81cb-2adb65dbc955]">https://dare.uva.nl/search?identifier=6ba47af3-8bf4-4b46-81cb-2adb65dbc955]</ext-link></mixed-citation>
			</ref>
			<ref id="R000038">
				<mixed-citation>Van Berkum, J.J., Van den Brink, D., Tesink, C.M., Kos, M., &#x0026; Hagoort, P. (2008). The neural integration of speaker and message. <italic>Journal of Cognitive Neuroscience</italic>, 20(4), 580–591. <ext-link xlink:href="https://doi.org/10.1162/jocn.2008.20054">https://doi.org/10.1162/jocn.2008.20054</ext-link></mixed-citation>
			</ref>
			<ref id="R000039">
				<mixed-citation>Van den Heuvel, H. (1996). <italic>Speaker variability in acoustic properties of Dutch phoneme realisations</italic>. Doctoral dissertation, Katholieke Universiteit Nijmegen. [<ext-link xlink:href="https://repository.ubn.ru.nl/bitstream/handle/2066/76416/76416.pdf">https://repository.ubn.ru.nl/bitstream/handle/2066/76416/76416.pdf</ext-link></mixed-citation>
			</ref>
			<ref id="R000040">
				<mixed-citation>Van Heuven, V.J. (1997). Effects of focus distribution and accentuation on the temporal and melodic organisation of word groups in Dutch. In S. Barbiers, J. Rooryck, &#x0026; J. van de Weijer (Eds.), <italic>Small words in the big picture. Squibs for Hans Bennis</italic>. HIL Occasional Papers no. 2 Leiden: Holland Institute of Generative Linguistics. 37–42.</mixed-citation>
			</ref>
			<ref id="R000041">
				<mixed-citation><person-group person-group-type="author"><string-name><surname>Voeten</surname>, <given-names>C.C.</given-names></string-name></person-group> (<date><year>2020</year></date>). <source>buildmer: Stepwise elimination and term reordering for mixed-effects regression</source>. <italic>R</italic> package version 1.5. <ext-link xlink:href="https://CRAN.R-project.org/package=buildmer">https://CRAN.R-project.org/package=buildmer</ext-link></mixed-citation>
			</ref>
		</ref-list>
	</back>
</article>