Coda glottalization is the process by which coda stops are produced either with simultaneous glottal constriction or with glottal constriction that has replaced the oral gesture. For example, in American English (as in many other varieties) the word bat /bæt/ may be pronounced without coda glottalization [bæt] (or with another non-glottal variant of /t/, such as [ɾ]), or with coda glottalization [bæʔ͡t, bæʔ]. In this study, we focus on instances of coda glottalization where the glottal constriction co-occurs with the oral one, e.g., /bæt/ → [bæʔ͡t]. Using a word recognition task with eye-tracking, we examine the timecourse of processing of coda glottalization in American English, to determine whether listeners use voice quality information anticipatorily to perceive voicing and place of articulation contrasts in coda stops.
1.1. Where, and how often, does glottalization occur in American English?
In this study, we use the term glottalization to refer to a phonological phenomenon characterized by increased vocal fold constriction associated with a particular segment. (In English and other languages, there are additional sources of vocal fold constriction, e.g., phrasal creak, which act on larger prosodic units). Phonetically, glottalization is usually realized as creaky voice, with or without a glottal stop [ʔ]. Different sources of glottalization exist, including word-initial glottalization (e.g., of words beginning with a stressed vowel: Dilley et al., 1996; Garellek, 2013; Davidson & Erker, 2014), and coda glottalization, which we focus on here.
Glottalized variants of coda /t/ include cases where /t/ is realized as a glottal stop with no alveolar closure (/t/ → [ʔ]), and cases where /t/ is ‘reinforced’ with glottal closure in addition to the alveolar closure (/t/ → [ʔ͡t̚, ʔ͡t]). The reinforced instances of /t/ are often unreleased (Esling et al., 2005), though released variants with glottalization, which we focus on here, are also attested (Seyfarth & Garellek, 2015; Penney et al., 2018). Altogether, glottalized variants of /t/ are common in American English, especially before sonorants (Huffman, 2005; Pierrehumbert, 1994; Sumner & Samuel, 2005; Seyfarth & Garellek, 2015). For instance, Seyfarth and Garellek (2015) show that speakers from the Buckeye Corpus (Pitt et al., 2007) glottalize coda /t/ as glottal stop [ʔ] roughly 50% of the time. Glottalized variants of coda /t/ in general (both [ʔ͡t] and [ʔ]) are more frequent than any other variant of /t/, including the canonical [t] variant. Previous studies have come to similar conclusions (Pierrehumbert, 1994; Huffman, 2005): In a corpus of Long Island English, Huffman (2005) reports that coda /t/ glottalizes 58% of the time.
Studies have also shown that coda /p/ in American English undergoes glottalization to [ʔ͡p] (but not to [ʔ]), though glottalization rates for /p/ are much lower than for /t/ (Pierrehumbert, 1994; Huffman, 2005; Seyfarth & Garellek, 2015). In Long Island English, glottalization of coda /p/ occurs roughly 29% of the time (Huffman, 2005). Thus, glottalized variant [ʔ͡p], though attested in American English, cannot be considered the most common variant of /p/ in coda position, in contrast to glottalized variants of /t/. Lastly, glottalized variants of other voiceless sounds (e.g., /k/ or /tʃ/) and voiced sounds (e.g., /d/) are not attested in American English, but can be found in varieties of British English (Roach, 1973; Foulkes & Docherty, 2007).
1.2. Glottalization as enhancement of voice and place contrasts
Several researchers have proposed that glottalized variants of coda stops are due to phonetic enhancement. One enhancement-based explanation for coda glottalization is that it occurs as a mechanism to inhibit voicing and/or to cue voicelessness (Stevens & Keyser, 1989; Pierrehumbert, 1994, 1995; Keyser & Stevens, 2006; Stevens & Keyser, 2010; Gordeeva & Scobbie, 2013; Penney et al., 2018). The reason why voiceless coda stops might involve glottalization in order to inhibit voicing is that these very sounds are likely to undergo voicing: Utterance-medially, stop voicing is favored aerodynamically (Westbury & Keating, 1986); coda stops are also often immediately preceded by a vowel, which has strong voicing, and are frequently followed by a voiced onset. For instance, it is well known that coda /t/ is more likely to glottalize before a sonorant (Pierrehumbert, 1994; Huffman, 2005; Keyser & Stevens, 2006; Seyfarth & Garellek, 2015). If the stop is glottalized, it is by definition produced with increased vocal fold constriction. This constriction should help cease voicing, ensuring that the voiceless coda stop is in fact voiceless.
However, the enhancement account of /t/ glottalization cannot explain why /t/ and /d/ often neutralize to a voiced tap intervocalically. Moreover, Huffman (2005) scrutinized this account on empirical grounds. She showed that coda glottalization was prevalent even before voiceless consonants, which is unexpected if coda glottalization should be used to enhance voicelessness. Moreover, Huffman found that phrase-final coda stops were not more likely to glottalize before (phrase-initial) sonorants than obstruents; assuming that phrase-initial obstruents have weaker voicing than phrase-initial sonorants, this finding is unexpected if /t/ and /p/ could stand to be enhanced before sonorants. Therefore, though glottalization is likelier to occur before sonorants than obstruents phrase-medially, this effect is weakened phrase-finally (see also Seyfarth & Garellek, 2015); more importantly—and contrary to an explanation rooted in enhancement of voicelessness—glottalization still occurs on coda stops that precede voiceless sounds.
As an alternative enhancement-based explanation, Keyser and Stevens (2006) further propose that all voiceless stops are likely to glottalize phrase-finally because voiced stops are likely to devoice in that position (see e.g., Westbury & Keating, 1986). To compensate for devoicing of voiced coda stops, voiceless stops undergo glottalization. Thus, according to this explanation, glottalization of phrase-final voiceless stops occurs not because voicelessness is difficult to produce or perceptually ‘imperiled’ in that position, but as a consequence of the fact that the voicing during voiced stops is weakened. Glottalization of voiceless stops thereby helps enhance the voicing contrast phrase-finally.
Note also that an explanation involving voiceless enhancement leaves open the question regarding distributional differences (i.e., in American English, why does coda /t/ glottalize so frequently, /p/ only sometimes, and /k/ rarely?). This issue is addressed by Stevens and Keyser (2010), who propose that coda glottalization also occurs to cue closure and release specifically for coda /t/. Coronal closure and release are more subject to gestural overlap than labial or velar constrictions for /p/ or /k/. Since overlap weakens recoverability of a particular feature or segment, /t/ glottalizes when the voiceless closure and coronal release would be obscured by a non-coronal (Keyser & Stevens, 2006; Stevens & Keyser, 2010). Seyfarth and Garellek (2015), however, show that glottalization rates for coda /t/ before non-coronals are roughly the same as rates of utterance-final /t/, for which there is no chance of overlap.
Overall it is clear that coda glottalization in American English is strongly associated with voiceless stops, and that it occurs more frequently with coda /t/ than /p/, and rarely with /k/. And while enhancement-based explanations for coda glottalization may not be able to account for all instances of the phenomenon, they dominate in the literature. In this paper, we test a central assumption of these enhancement-based accounts, namely, that glottalization (as an enhancement gesture/feature) is perceptually beneficial to listeners. If so, we expect that listeners should be more likely to recognize words with coda /t/ and /p/ when they are glottalized. In other words, the presence of glottalization should result in faster and better recognition of words ending in voiceless stops, especially /t/.
1.3. Perception of phonological variants and coarticulation
In addition to testing predictions of enhancement theories, this study has implications for understanding how phonological variants are perceived. Several researchers have examined listeners’ ability to recognize familiar words produced with a glottal variant of /t/ both in word-medial (Pitt et al., 2011) as well as in word-final coda position (Sumner & Samuel, 2005). In particular, word-final glottal variants of /t/ (both [ʔ͡t or ʔ]) were as effective as canonical non-glottalized [t] at priming a semantically-related target. This provides evidence that glottalization does not hinder word recognition. Although glottalization does not hinder recognition of coda /t/, its effect on coda /p/ and voiced stops, however, is yet to be determined.
Glottalization is phonologically associated with a coda stop, but it is additionally (or even entirely) realized phonetically as creaky voice on the preceding vowel (Pierrehumbert, 1994; Huffman, 2005; Garellek, 2015; Seyfarth & Garellek, 2015). Thus, a word like bat, when glottalized, may be realized as [bæ̰t, bæ̰]. Listeners might therefore use creaky voice (derived from glottalization) as a cue to anticipate the upcoming coda consonant. In the current study, we investigate this question using a variant of the visual world paradigm (Allopenna et al., 1998), where listeners’ eye movements on a visual display are continuously monitored while they are listening to speech stimuli. This paradigm allows us to assess listeners’ interpretation, shown through their visual fixations, of the incoming speech signal, and how this changes over time as the signal unfolds.
A number of studies have used eye tracking to assess listeners’ perception of coarticulated speech. For example, Dahan et al. (2001) cross-spliced the initial CV of minimal pairs of words such as neck and net to create tokens which either contained matching coarticulatory cues to the final coda consonant (e.g., [nɛ] in net cross-spliced with [t] from another token of net) or mismatching cues (e.g., [nɛ] in neck cross-spliced with [t] from net). Although the mismatching tokens were eventually perceived as net, the formant transition cues in the vowel were more consistent with neck. Listeners, in fact, made more initial fixations to the competitor image of a neck when they heard net tokens with mismatch cues than when they heard tokens with matching cues (see also Dahan & Tanenhaus, 2004 on Dutch and Gow & McMurray, 2007 on C-C coarticulation in English). Thus listeners make use of the information present in the speech signal immediately, and their interpretations are incrementally updated as the speech signal unfolds. More importantly, the inappropriateness of a particularly coarticulatory cue can hinder the recognition process by garden-pathing listeners’ initial interpretations.
In fact, listeners are not just sensitive to the appropriateness of coarticulatory cues for a given upcoming segment, but are also sensitive to the specific timing of when coarticulatory cues become available in the speech signal. Beddor et al. (2013) investigated the timecourse of listeners’ perception of coarticulatory vowel nasalization in American English. Participants heard cross-spliced target words (e.g., bent) with nasalization starting early or late in the vowel while looking at two pictures on a display. Listeners’ fixations converged on the target more quickly when nasalization started early in the vowel than when it started late in the vowel. These results thus show that lexical access follows closely the unfolding of acoustic cues in the speech signal (see also Salverda et al., 2014). Given the visual world paradigm’s (Allopenna et al., 1998) success at elucidating listeners’ online perception of coarticulatory cues, in this study, we use this paradigm to examine listeners’ online perception of coda glottalization in American English.
1.4. Hypotheses of the present study
The previous work discussed above has shown that both the frequency of phonological variants and the coarticulatory information they provide can influence online word recognition. In the case of glottalization, we therefore expect that listeners should show online sensitivity to its acoustic presence, and that they should have prior expectations of which sounds are likelier than others to be glottalized. The research on phonetic enhancement also makes clear (but not yet verified) predictions regarding how listeners should perceive acoustic information deemed ‘enhancing’: If an enhancement feature is present, the sound it enhances should be perceived more rapidly and identified more accurately. If the presence of glottalization is in fact due to enhancement (either for /t/ specifically, or for voiceless stops more generally), we expect listeners to recognize a target word faster and/or better when it is present than when it is absent.
In sum, we examine here two hypotheses related to coda glottalization in American English. First, since glottalization occurs more frequently with /t/ than with /p/, glottalization should preferentially facilitate recognition of /t/ words more than /p/ words. Second, if glottalization is associated with voiceless stops more generally, as Keyser and Stevens (2006) suggest, then glottalization should facilitate recognition of voiceless stops (both /t/ and /p/) over voiced stops.
In this study, we use eye-tracking to measure online word recognition using visual fixations to printed words. In brief, eye movements are tracked as participants hear instructions to interact with a display, such as: “Look at the word bat.” The display contains two words, e.g., the target bat and a competitor like bad, represented as printed words. Following previous researchers (e.g., McQueen & Viebahn, 2007; Mitterer, 2011; Brouwer et al., 2012), we chose to use the printed-word variant of the visual world paradigm (rather than the traditional variant with images of objects), because of difficulty representing certain targets visually. Crucially, the target and other words presented on the screen differed minimally in their orthography; for example, English bat was paired with bad.
Sixty participants (20 male, 46 female; Mean age: 20.56, range: 18–38) were recruited from the UCLA Psychology Subject Pool and received course credit in an undergraduate psychology or linguistics class. Participants were all adult native speakers of American English without any reported hearing deficits. All had normal or corrected-to-normal vision and could recognize the visual stimuli used. An additional 13 participants were also tested, but their data were excluded from data analysis because they were not native English speakers (n = 7), because of inattention during the task (n = 4), or because of failure of the system to save the experimental data (n = 2).
2.2. Target words
The target words were monosyllabic CVC English words. Words were chosen such that they had a minimal pair differing in terms of the coda obstruent. Pairs of words were also controlled for orthographic length and were organized into four groups, each consisting of four word pairs.
2.2.1. The Baseline group
In the first Baseline group, we tested whether the presence of glottalization resulted in differences in eye-tracking behavior even when glottalization is not associated with a regular glottalized variant of a coda. Thus for this condition we used words differing only in their final consonant, which could be either a voiceless alveolar fricative /s/ or a voiceless dental fricative /θ/, e.g., moss vs. moth. Neither fricative is regularly glottalized in coda position. The complete wordlist is shown in Table 1 (the pairs do not differ significantly in frequency, as assessed by a paired t-test).
|Word 1||Log frequency||Word 2||Log frequency|
For these words, the initial consonant and vowel preceding the final fricative were always drawn from the same original token (always the alveolar, e.g., moss), to minimize the acoustic differences between words in these pairs. Although in natural speech, formant transitions before alveolar vs. dental fricatives differ slightly (Jongman et al., 1985), our pilot experiments revealed that both words in a pair (e.g., moss and moth) sounded equally natural when resynthesized with the same pre-alveolar vowel. Overall then, for this group we predict no overall facilitative or inhibitory effect of glottalization, because (1) neither glottalized [ʔ͡s] nor [ʔ͡θ] are regular variants of coda /s/ and /θ/ in American English, and (2) both fricatives are voiceless, which precludes the possibility of voicing enhancement. For example, if on a particular trial, listeners hear the word mass and see both MASS and MATH on the screen, the presence of glottalization should not affect their identification of the target given the alternative on the screen; although the presence of glottalization might be unexpected on mass, it is not expected on the competitor MOTH either, and thus should not bias the listener towards the competitor, resulting in an overall null effect of glottalization.
2.2.2. The Place of Articulation (POA) group
The second Place of Articulation (POA) group tested whether the presence of glottalization would result in differences in eye-tracking behavior for words ending in coda /t/ vs. /p/. For this condition we used words differing only in their final consonant, which could be either a voiceless alveolar stop /t/ or a voiceless labial stop /p/, e.g., rap vs. rat. In American English, coda /t/ usually appears as glottalized [ʔ͡t] or even [ʔ], whereas coda /p/ glottalizes much less frequently (Pierrehumbert, 1994; Huffman, 2005; Seyfarth & Garellek, 2015). Thus, for this group we have two main predictions: If glottalization is associated with /t/ only, then its presence may be facilitative or neutral for identifying words with /t/ (vs. /p/), but not inhibitory. Further, its presence on a word with coda /p/ should be inhibitory when paired with a competitor ending in /t/. On the other hand, if listeners associate glottalization with both /t/ and /p/ (consistent with the fact that glottalization can occur with both), then we would expect no inhibitory effect of glottalization on identifying words with /p/. The complete wordlist for the POA group appears in Table 2 (the pairs do not differ significantly in frequency, as assessed by a paired t-test).
|Word 1||Log frequency||Word 2||Log frequency|
2.2.3 The Voicing groups
The third and fourth Voicing groups tested whether the presence of glottalization resulted in differences in visual fixations for words ending in voiceless vs. voiced coda stops. Thus for these conditions we used words differing only in their final consonant: In Group 4a, the final stop could be either a voiceless alveolar stop /t/ or a voiced /d/ (e.g., bat vs. bad); in Group 4b, the final stop could be either a voiceless labial stop /p/ or a voiced /b/ (e.g., tap vs. tab). Recall that, in American English, only the voiceless codas glottalize, though coda /p/ glottalizes much less frequently than /t/ (Pierrehumbert, 1994; Huffman, 2005; Seyfarth & Garellek, 2015). However, some researchers (Pierrehumbert, 1994; Keyser & Stevens, 2006) have claimed that glottalization occurs to enhance voicelessness, which should in principle apply to both /t/ and /p/. Thus, for this group we have two main predictions: If glottalization enhances voicelessness, then its presence should be facilitative for words with coda /t, p/ and inhibitory for /d, b/. Alternatively, if glottalization is associated with /t/ more than /p/, then its presence may be facilitative for words with /t/, weakly facilitative or neutral for identifying words with /p/, and inhibitory for words with coda /d, b/. The complete wordlists for Groups 4a and 4b are shown in Table 3 (the pairs do not differ significantly in frequency, as assessed by a paired t-test).
|Voiceless||Log frequency||Voiced||Log frequency|
In total then, we had 16 pairs of target words. Correspondingly, we included 16 pairs of filler words, which differed in terms of their initial onset consonant. The reason for using pairs differing only in their onset was to prevent participants from learning that they should focus only on the final consonant. As with the target word pairs, these words were always monosyllabic and came in non-glottalized/glottalized versions. A complete wordlist (including fillers) can be found in the Appendix.
2.3. Audio stimuli resynthesis
The stimuli were recorded by a phonetically-trained female speaker of Southern Californian English. She uttered each word in a carrier phrase, where the target word appeared in medial position to avoid phrase-final creak. Each word was then resynthesized using the Klatt synthesizer in Praat (Boersma & Weenink, 2015) as follows. First, natural tokens were segmented according to their onsets, vowels, and codas. (For words beginning with approximants /ɹ/ and /w/, the onset was not segmented due to difficulties identifying onset-vowel boundaries, and because such approximants can easily be resynthesized to sound natural.) The natural onsets and codas were extracted. Then, the vowel was extracted using a Hamming window (in our piloting, windowing improved the naturalness of the vowel resynthesis and consonant-vowel cross-splicing, described below). The vowel was then resynthesized using the Klatt synthesizer in Praat. We created two versions of each vowel: One glottalized (i.e., with creaky voice during the vowel), the other non-glottalized. The four parameters used to create both versions are shown in Table 4, and result in a prototypical creaky voice quality (Keating et al., 2015; Garellek, to appear) that is commonly found on vowels before glottalized /t/: Constricted-sounding (as indexed by lower spectral tilt), low and irregular in F0, and where the creaky voice progresses over the vowel’s duration (Garellek & Seyfarth, 2016). The precise values of each parameter, and the time points at which values were changed, were determined in our piloting to maximize the percept of glottalization while minimizing synthetic vowel quality.
|Non-glottalized vowel||Glottalized vowel|
|Open Phase (OQ)||0.5 throughout||Initial value: 0.5
Middle value: 0.5
Final value: 0.2
|Spectral tilt||10 throughout||Initial value: 10
Final value: 0
|Flutter||0 throughout||0.95 throughout|
|Double pulsing||0 throughout||Initial value: 0
Middle value: 0
Final value: 0.5
The first parameter, Open Phase, changes the open quotient of each glottal pulse (the proportion of the glottal cycle during which the vocal folds are open); higher values of Open Phase result in a more prominent first harmonic, which is associated with less creaky voice. We chose to manipulate this parameter from the vowel’s midpoint rather than onset, because during piloting we found that a low level of Open Phase throughout the vowel resulted in a less natural-sounding glottalized token with too tense a voice quality. Thus, for our glottalized vowels, we had this parameter decrease linearly from the midpoint of the vowel to its end. This resulted in an increasingly constricted, creaky quality of the glottalized vowel compared to its non-glottalized counterpart.
We also manipulated the Spectral Tilt parameter, which raises or lowers the spectral tilt between H1 and the harmonic around 3000 Hz. Glottalized vowels typically have lower spectral tilt, especially in the latter half of the vowel (Garellek & Seyfarth, 2016). Consequently, we had the vowel start off with a Spectral Tilt value of 10 dB and end with a value of 0 dB. This parameter, like the Open Phase one, also resulted in a more constricted creaky voice quality for glottalized vowels compared with non-glottalized ones.
Glottalized tokens also had adjustments to the Flutter parameter, which at higher values produces jitter, or irregular pitch periods. Higher values of the parameter mean that the F0 becomes increasingly irregular, which is common during glottalization (Garellek & Seyfarth, 2016). The non-glottalized vowels had no Flutter, resulting in a regular falling F0 contour.
The final parameter manipulated was Double Pulsing, which like Flutter varied for glottalized tokens and held at 0 for non-glottalized ones. A non-zero value of this parameter creates sub-harmonics, resulting in period-doubling or ‘diplophonia,’ another common acoustic and perceptual feature of glottalization and creaky voice in general (Gerratt & Kreiman, 2001; Redi & Shattuck-Hufnagel, 2001; Keating et al., 2015). Higher values indicate greater period doubling, increasing the percept of irregular pitch. We chose to manipulate this parameter from the vowel’s midpoint rather than onset, because during piloting we found that a high level of Double Pulsing throughout the vowel resulted in a less natural-sounding glottalized token. Thus in our glottalized vowels, this parameter increased from 0 to 0.5 from the vowel’s midpoint to the end of the vowel.
Once the non-glottalized and glottalized versions of each vowel were resynthesized, we then spliced the original (non-synthesized) onset and coda back on to the synthetic vowels. All tokens had released codas, though these could differ in terms of intensity across items. If the onset or coda was judged too loud or quiet with respect to the synthetic vowel, we adjusted the consonants’ intensities before concatenating the two versions of the word once more. The vowel durations—which were not altered in the resynthesis—could differ across items, based on the way the speaker uttered each word. As we would expect, vowels tended to be shorter before voiceless stops than before voiced ones, and before coronals than before labials (Peterson & Lehiste, 1960; Luce & Charles-Luce, 1985). Crucially, non-glottalized and glottalized versions of each word always had the same onset and coda, and the same vowel duration. Lastly, all stimuli were scaled for peak intensity. Figure 1 illustrates the spectrographic differences between non-glottalized and glottalized versions of the audio stimuli. Sample audio files can be found as supplementary materials on https://www.adamjchong.com/publications.html.
2.4. Visual stimuli
Visual stimuli consisted of printed text of the target words. These were created in Adobe Photoshop and were 88 pixels in size and in Times New Roman font. They were saved as .bmp files and were presented on a 1920 × 1080 ASUS HDMI monitor. A sample visual display is shown in Figure 2.
Each listener heard each target test word only once, in either glottalized or non-glottalized conditions. Listeners, however, heard both members of a given stimulus pair across two different blocks, such that a particular visual stimulus pair was not seen in the same block. The order of blocking was also counterbalanced. This yielded four experimental lists. Listeners heard onset fillers in both glottalized and non-glottalized conditions in all counterbalancing groups, split evenly across blocks. The blocking for these stimuli was kept consistent across each experimental list, with trials in each block being fully randomized. There were 96 trials in each experimental list (see Appendix for details).
Participants sat in front of an arm-mounted SR Eyelink 1000 (SR Research, Mississauga, Canada) set to track the left eye at a sampling rate of 250 Hz, at an approximate distance of 550 mm. At the start of each experiment, a 5-point calibration was conducted. They were instructed that they were going to see words on a screen and to listen very carefully to the instructions they were going to hear. At the start of each trial, both words appeared on the screen, one on the left and the other on the right, and they heard the instructions: “Look at the words on the screen.” The text remained on the screen for 3 seconds. Then they heard the instructions: “Now look at the cross,” which coincided with the appearance of a cross in the center of the display. This was done to ensure that when the words were named in the test phase, listeners were not already on the relevant target word. Following a 500-ms pause, they then heard: “Now look at the word…[TARGET].” The text stayed on the screen for 1500 ms following the offset of the target word. The side that the target image appeared on was counterbalanced across trials, such that they appeared equally on either side of the screen. In order to familiarize participants with the task, listeners heard six practice trials which consisted entirely of onset fillers (which also appeared later in the experiment itself). Drift correction was conducted after every five trials. In total, each experimental session lasted approximately 25 minutes. The experiment is schematized in Figure 3.
The eye movements of participants were monitored over the course of each test trial until the trial ended. We were interested in looking at two measures, following Beddor et al. (2013). The first was the latency of the first correct fixation to the target word. Latencies were calculated from the onset of the vowel. We chose the vowel onset as the start of the analysis window (or approximant onsets, which were also resynthesized) as we are interested in within-item comparisons (e.g., non-glottalized bat vs. glottalized bat), where the phonetic differences between the resynthesized stimuli (i.e., presence of creaky voice) is realized from vowel (or approximant) onset to vowel offset. Because words are matched for non-approximant onsets, any disambiguating information would only start during the vowel via formant transitions. Trials in which listeners made errors in looks (i.e., they only looked to the distractor and never to the target) were excluded from this analysis. We excluded any fixations that occurred in the first 200 ms from the vowel onset. Given the 200 ms it takes to initiate a saccade (Matin et al., 1993; Dahan et al., 2001), any fixation to the target region would not be due to any potentially disambiguating information. We further excluded trials in which the latency of the first correct fixation did not occur within the first 1000 ms. Two rectangular interest areas were set 350 × 240 pixels around the printed target words, and were situated in distinct halves of the screen such that there was no overlap between interest areas. The interest areas were also larger than the target words themselves to ensure that fixations around the target word would also be counted.
The second measure was the proportion of fixations to the target image, which indicates how good a target auditory stimulus is at activating the target lexical representation. Thus if an image attracts more looks than a competitor when one auditory stimulus is heard compared to another, we can say that the former is better at facilitating lexical access than the latter. The time window for the analysis started from 200 ms following the onset of the vowel and ended at 1000 ms, where fixations to the target image generally plateaued. This resulted in an 800-ms analysis window. Finally, since the analysis of proportions is problematic to handle in a linear model, target fixation proportion data was first transformed using the empirical logit function, using the function in Barr (2008).1
To ensure that our synthesized glottalized stimuli were not degraded compared to our synthesized non-glottalized stimuli, we first examine what listeners’ responses were to these stimuli in our baseline condition where we do not expect there to be any differences in fixation latencies or proportions based on presence or absence of glottalization. Thus in the baseline conditions, trials consisted of pairs of target words with coda /s/ or /θ/: Glottalization is not associated with either of these coda consonants, so we do not expect any facilitatory or inhibitory effect on word recognition in this sub-experiment.
The left panel of Figure 4 shows the log-transformed mean latencies of the first correct fixations on the trials where participants heard trials with coda /s/ or /θ/. Overall, as we predicted, listeners fixated to the target word equally quickly regardless of whether they had heard a glottalized or non-glottalized token of the target word (Glottalized: Mean = 587, SD = 152; Non-glottalized: Mean = 579, SD = 169). A linear mixed-effects model (Bates et al., 2015) with Satterthwaite approximations was fit using the lmerTest() function (Kuznetsova et al., 2016) in R (R Core Team, 2015) with the log-transformed fixation latencies as the dependent variable and Glottalization (Glottalized vs. Non-glottalized) as the fixed effect. The model was also fit with random intercepts for participant and target word, as well as random slopes for Glottalization status for both participant and word. P-values are provided in the model outputs. (For all statistical analyses, the full model structures and results can be found as supplementary materials on https://www.adamjchong.com/publications.html). As was expected, the presence or absence of glottalization did not significantly affect the speed at which listeners looked to the target word when they heard a word with coda /s/ or /θ/ [β = –0.07, SE = 0.09, t = –0.73, p > 0.05].
Further confirmation of this result comes from looking at the timecourse of fixations to the target word in baseline trials (right panel of Figure 4). As is evident from the fixation behavior, fixations to the target word did not differ across time regardless of the presence or absence of glottalization in the auditory stimuli. A linear-mixed effects model was fit to the logit-transformed target fixation proportion, with Glottalization as a fixed effect as well as random intercepts for participant and word as well as random slopes for Glottalization. Fixations were averaged over the 800-ms analysis window that started from 200 ms following the onset of the vowel and ended 1000 ms after the vowel onset. As with the analysis of fixation latencies, there was no significant effect of glottalization on listeners’ fixation proportions for target words with coda /s/ and /θ/ [β = –0.41, SE = 0.26, t = –1.62, p > 0.05]. As expected, the results of both the latency and fixation analysis indicate no specific baseline effects of glottalization on the resynthesized words.
3.2. Place of Articulation
In the Place of Articulation condition, listeners saw displays with minimal pairs of words with coda /t/ or /p/. Here we were interested in examining whether listeners were faster at recognizing /t/ and /p/ words when they were glottalized.
Table 5 and Figure 5 show the mean latencies of the first correct fixations on the trials where participants heard trials with coda /t/ or /p/. Log latencies were analyzed using a linear-mixed effects model with Glottalization and Place (both contrast-coded), as well as their interaction, as fixed factors. The model also included by-subject and by-item random intercepts as well as random slopes for Glottalization X Place by subject and a random slope for Glottalization by target word. There was no significant effect of Glottalization [β = 0.13, SE = 0.11, t = 1.20, p > 0.05] or Place [β = 0.29, SE = 0.18, t = 1.62, p > 0.05]. Finally, there was a marginal interaction, indicating that listeners were marginally quicker to fixate on the target when they heard a word ending in a glottalized /t/ than one ending in a non-glottalized /t/ [β = –0.36, SE = 0.21, t = –1.77, p = 0.09].
|Coronal stops||Labial Stops|
|Glottalized||491 (121)||561 (142)|
|Non-Glottalized||532 (125)||557 (152)|
The timecourse of fixations to the target and distractor is shown in Figure 6. We analyzed fixation proportions using the same analysis window as above and with a linear-mixed effects model. For this model, we included as fixed effects Glottalization and Place, as well as their interaction. Random intercepts for subject and target word were also included, as well as random slopes for Glottalization and Place by subject and Glottalization by word. This was the maximal model to converge. Results show no significant effect of glottalization [β = 0.04, SE = 0.21, t = 0.18, p > 0.05] but there was a significant effect of place [β = –1.56, SE = 0.56, t = –2.79, p = 0.03], confirming the observations that words with coda /p/ were just recognized more poorly than words with coda /t/. The interaction was not significant [β = –0.02, SE = 0.42, t = –0.04, p > 0.05], indicating that the effect of glottalization did not differ between words ending in /p/ vs. /t/.
Overall then, listeners looked less to the correct target when they heard a word ending in coda /p/ than a word ending in coda /t/, suggesting that listeners expect a word-final /t/ more than a word-final /p/. This may be due to distributional properties of English: In the SUBTLEXus corpus (Brysbaert & New, 2009), there are 2.5 times more words ending in /t/ than /p/. That is, it is possible listeners might a priori expect a /t/-final word than /p/-final word when presented with both. One could also argue that this effect is driven by acoustic properties, for instance that coda /p/ is harder to identify than /t/. However, this is unlikely to be due to properties of the /p/-final stimuli independent of the distractor; as we will see in the following section, when the target word ends in /p/ and the competitor ends in /b/, listeners do in fact converge on /p/ responses by the end of the analysis window.
In the Voicing condition, listeners saw displays with minimal pairs of /t/ vs. /d/, as well as /p/ vs. /b/. Here we address whether or not glottalization facilitates the recognition of voiceless stops over voiced stops, and further whether this differs by place of articulation.
Figure 7 shows the log-transformed mean latencies of first correct fixations in the Voicing condition. Overall listeners were slower at correctly fixating to the target image when the target word ended in a voiced stop (Table 6). A linear mixed-effects model was fit to fixation latencies with the following fixed effects: Voicing, Place, Glottalization, as well as their interactions. All factors were contrast-coded. Random intercepts for subject and target word were also included, as well as by-subject random slopes for Glottalization X Place, Glottalization X Voicing, and Voicing X Place, as well as by-word slopes for Glottalization. This was the maximal model to converge. There was a significant effect of Voicing [β = 0.60, SE = 0.08, t = 7.60, p < 0.001], with listeners being slower at fixating to the correct target image when the target was a voiced stop, regardless of place of articulation or glottalization. There was also a significant two-way interaction of Glottalization by Voicing [β = –0.28, SE = 0.12, t = –2.30, p = 0.03], with listeners being slower at fixating to the correct image when the target word was voiced and glottalized. No other effects were significant.
|Coronal stops||Labial stops|
|Glottalized||493 (121)||607 (138)||489 (149)||607 (139)|
|Non-Glottalized||506 (137)||590 (151)||507 (138)||562 (135)|
The timecourse of fixations to the target and distractor is shown in Figure 8. Logit-transformed fixation proportions were analyzed using a linear-mixed effects model with the following fixed effects: Voicing, Place, Glottalization, as well as their interactions. All factors were contrast coded. The random-effects structure was the same as for fixation latency. This was the maximal model to successfully converge. Results show significant main effects of Voicing [β = –0.66, SE = 0.17, t = –3.86, p = 0.002] and Glottalization [β = 0.33, SE = 0.13, t = 2.61, p = 0.01]. There was also a significant two-way interaction of Voicing by Glottalization [β = 0.87, SE = 0.27, t =3.21, p = 0.003], driven by the fact the non-glottalized voiced stops have higher fixation proportions than glottalized voiced stops (but no effect of glottalization is found for voiceless stops, as we discuss in more detail below). No other factors and interactions were significant.
Given the significant interaction between Glottalization and Voicing, we were further interested in examining when fixations to words ending in non-glottalized stops were higher than those to words ending in glottalized ones. (No timecourse analyses were presented earlier for lack of an effect of glottalization). We used the bdots package (Seedorff et al., 2017) in R to ascertain the time window in which fixations to the target word differed as a function of glottalization. Following Oleson et al. (2015), a nonlinear curve is fit to the fixation proportions for each target word2 in the glottalized and non-glottalized conditions. In our case, since we did not find a main effect of place, both /t/ and /p/, as well as /d/ and /b/ are collapsed into one analysis per voicing group. Through bootstrapping using the logistic.boot() in the bdots package function in R, a set of estimates and standard errors is obtained across all the items within a glottalization condition. This then yields a time window in which two curves (e.g., glottalized vs. non-glottalized voiceless stops; glottalized vs. non-glottalized voiced stops) significantly differ from each other.
Our analysis revealed no time windows in which participants looked significantly more at the target word when glottalization was present for words ending in voiceless stops. For voiced stops on the other hand, participants looked significantly more at the target word when they heard the non-glottalized production starting from 520 ms following the onset of the vowel and lasting to the end of the trial (adjusted p = 0.006, see also figure in supplementary materials on https://www.adamjchong.com/publications.html). Allowing for the 200 ms time delay to execute a saccade in response to an auditory stimulus, this means that participants are responding to what they heard at about 320 ms following the vowel onset, which is before the average onset of the coda consonant (400 ms), but nonetheless within the second half the vowel where glottalization would be most prominent. This therefore provides further evidence that participants’ fixation behavior was due to their perception of the glottalization on the vowel, particularly when the strength of glottalization increased.
4.1. Summary of results
Using a printed-text variant of the visual world paradigm, we investigated the timecourse of perception of coda glottalization in American English. Specifically, we examined the predictions of two enhancement accounts for why glottalization occurs. First, several researchers (e.g., Stevens & Keyser, 1989; Pierrehumbert, 1994, 1995; Keyser & Stevens, 2006; Stevens & Keyser, 2010; Gordeeva & Scobbie, 2013; Penney et al., 2018) have claimed that glottalization enhances voicelessness; thus, we predicted that glottalization should facilitate recognition of voiceless stops (both /t/ and /p/) when paired with voiced ones. When the target word ended in a voiceless stop, we found that listeners fixated to the target image equally quickly, regardless of the absence or presence of glottalization. That is, glottalization did not facilitate the recognition of words with coda voiceless stops. Interestingly, we found that glottalization inhibited the recognition of words with coda voiced stops, as evidenced by slower target fixation latencies as well as lower target fixation proportions.
Second, glottalization most commonly occurs with /t/ (Pierrehumbert, 1994; Huffman, 2005; Seyfarth & Garellek, 2015). Keyser and Stevens (2006) explain that this is due to the fact that coronal articulations are fastest to produce, which means that /t/ is likely to be subject to more overlap with non-coronal articulations and would therefore benefit from glottalization. Thus, we hypothesized that glottalization should facilitate the recognition of words ending in coda /t/ when paired with words ending in coda /p/. Contrary to this prediction, listeners were not any better at recognizing the target /t/-final word when it had glottalization. The same result was found for /p/-final words. Thus, although glottalization most commonly occurs with /t/ and less so with /p/, this distributional asymmetry does not result in faster or better recognition of words ending in /t/, nor does it result in slower or poorer recognition of words ending in /p/.
Overall, we interpret these results as follows: First, given that recognition of words ending in voiced stops is inhibited by glottalization, listeners must (a) be sensitive to its presence in the acoustic signal, and (b) know that glottalization is a cue to word-final voiceless stops but not to word-final voiced ones. Second, because listeners do not use glottalization to facilitate recognition of words ending in voiceless stops, this suggests that glottalization, though perceptible, does not enhance the recognition of words ending in voiceless stops in general, or /t/ specifically. In the remainder of the discussion, we review the implications of these results for the perception of coarticulation and explanatory theories of coda glottalization.
4.2. Glottalization as coarticulation
As we discussed earlier, listeners are sensitive both to the appropriateness of coarticulatory cues for an upcoming sound (Dahan et al., 2001; Dahan & Tanenhaus, 2004; Gow & McMurray, 2007), as well as to the specific timing of when coarticulatory cues become available in the speech signal (Beddor et al., 2013; Salverda et al., 2014). For instance, Beddor et al. (2013) showed that listeners’ fixations converged on a CVNC word (e.g., bent) more quickly when nasalization started early in the vowel than when it started late in the vowel.
In our study, creaky voice can also be construed as a coarticulatory cue to stop identity: If listeners know, based on distributional evidence, that the presence of creaky voice is a cue to an upcoming voiceless coda stop (especially /t/), then they should be able to use the non-modal phonation as an anticipatory cue to recover an upcoming glottalized stop. In other words, they should look faster at words ending in /t/ than /p/ when glottalized, and faster at words ending in /t/ or /p/ than /d/ or /b/. However, we find little support for this in our study: When they hear a glottalized word ending in a voiced stop, listeners fixate less on the target image but when they hear a glottalized word ending in a voiceless stop, they do not use glottalization to fixate sooner on the target.
Thus, our results seem to be at odds with those in previous work, particularly those in Beddor et al. (2013). We do not believe that the absence of an effect of glottalization in certain conditions can be attributable to a lack of statistical power. While Beddor et al.’s study had many more trials per participant (360 vs. 96), the current study tested nearly three times the number of participants (23 vs. 60). Moreover, the number of stimuli per condition was comparable across both studies. Instead, we believe that the absence of an effect of glottalization stems from intrinsic differences between cues to nasalization vs. glottalization. In our study, glottalized stimuli became progressively creakier as the vowel unfolded; two of the four glottalization parameters were used only from the vowel midpoint onward. We created glottalized stimuli in this way because of ecological validity and for methodological reasons: Vowels before glottalized stops are creakiest nearest to the stop formation—unlike vowels in phrasal creak, where the creaky voice quality remains more stable (Garellek & Seyfarth, 2016). Thus, we created stimuli that sounded typically glottalized (i.e., typical of vowels before glottalized stops), and which were unlikely to sound like stimuli produced with an overall creaky voice setting.
On the other hand, nasalized vowels in Beddor et al. (2013) varied according to where nasalization began (early vs. late onset), which the authors manipulated by varying the point at which oral vowels were cross-spliced with nasalized ones. Although it is unclear how the degree of nasalization (i.e., its spatial distinction with respect to oral vowels) varied over time, the results of their study imply that acoustic cues to nasalization must be perceptible as soon as they begin. This may be due to differences in listener sensitivity to glottalization vs. nasalization cues; for instance, listeners may be worse at perceiving changes associated with glottalization (including spectral tilt and F0 regularity) than they are at perceiving changes associated with nasalization (including formant frequencies and bandwidths, as well as spectral slope: Beddor, 1993; Styler, 2017). An alternative explanation is that, regardless of how well listeners perceive certain acoustic parameters, the cues to nasalization are more closely associated with the linguistic nasal gesture than the acoustic cues to glottalization are with stop identity. This is likely the case, given that cues to nasalization relate directly back to nasal gestures (nasal voice is a cue to the presence of a nasal gesture), whereas cues to glottalization relate only indirectly back to a coronal or labial gesture (creaky voice is a cue to the presence of a glottal constriction gesture, which in turn cues the presence of a voiceless coda stop). Moreover, glottal constriction gestures in English can be attributed to multiple linguistic phenomena, including coda stop glottalization, glottalization of word-initial vowels, hiatus environments, and phrasal creak (Dilley et al., 1996; Garellek, 2013; Davidson & Erker, 2014; Garellek, 2015); in the case of coda stop glottalization, glottal constriction is also optional, unlike vowel nasalization before nasal consonants, which is nearly categorically present (though variable in degree; Zellou, 2017). So differences in the online perception of nasal coarticulation (Beddor et al., 2013) vs. glottal coarticulation in our study might stem from the fact that perceiving glottal constriction is only indirectly tied to retrieving coda stop gestures, that glottal constriction is optional, and that it may also be attributed to other linguistic sources.
4.3. Glottalization as phonetic enhancement and phonological variation
A central goal of this study was to examine the validity of a critical assumption of listener-based enhancement theories for glottalization: Namely, that listeners should be able to perceive and utilize enhancing cues. Using a paradigm designed to probe listeners’ online perception of the incoming signal, we found no evidence in support of this assumption. On the other hand, our findings support Huffman (2005), who argued that enhancement alone could not account for all instances of coda glottalization. Huffman instead posited that coda glottalization might be an optional gesture associated with /t/ and /p/ in coda position. Although listener-based enhancement accounts of glottalization cannot account for our results here, it still is possible that glottalization is used to enhance the articulation of coda stops /t/ and /p/. In other words, speakers may produce glottalization to facilitate reaching an articulatory target (namely voicelessness), but that the glottalization gesture itself does not necessarily aid the listener. Note that, if this is the case, it is still unclear why coda /k/ would not also benefit from glottalization as a voicelessness enhancement strategy. Keyser and Stevens (2006, p. 43) suggest that the tongue surface stiffens during coda /p/ and /k/, which accelerates the intraoral pressure buildup and thus facilitates voicelessness, obviating the need for glottalization. Of course, that in turn does not explain why coda /p/ is sometimes glottalized, though (following the logic in Keyser & Stevens, 2006; Stevens & Keyser, 2010) perhaps it is subject to more articulatory overlap than /k/. The hypothetical need for enhancing coda /t/ in English is all the more puzzling, given that it has low informativity and is generally subject to the most patterns of lenition (Cohen Priva, 2017).
Nonetheless, the possibility that glottalization is an articulatory enhancement strategy could be viewed as similar to the role of creaky voice on the low-falling tone in White Hmong; though speakers reliably produce this tone with creaky voice (Esposito, 2012; Garellek, 2012), listeners ignore the creaky voice during word identification (Garellek et al., 2013). In that study, the authors hypothesized that creaky voice might be a means of achieving a low pitch, even if listeners do not use it independently of F0. Similarly, in this study we find that glottalization is not used by listeners to perceive voiceless stops, though coda stops (especially /t/) are usually glottalized in spontaneous speech (Huffman, 2005; Seyfarth & Garellek, 2015). Thus, glottalization might enhance the articulation of voicelessness and/or /t/, but this additional gesture is not being used by listeners as an enhancing cue.
Although we did not find evidence in support of listener-based enhancement as an explanation for the occurrence of coda stop glottalization, our results do in fact accord with studies of recognition of phonological variants. Firstly, we found that words with glottalized coda /t/ are recognized as well as those with canonical non-glottalized codas, which replicates the findings of Sumner and Samuel (2005) that both canonical non-glottalized coda [t] as well as glottalized coda [ʔ͡t] are equally effective in priming a given target word. Our current study both replicates this main finding using a different paradigm, and also extends the findings to the recognition of glottalized [ʔ͡p] variants, which Sumner and Samuel (2005) did not examine. Unlike glottalized [ʔ͡t], glottalized [ʔ͡p] is not the most frequent variant of /p/ in coda position (Huffman, 2005). Interestingly, we found that [ʔ͡p] is just as effective at facilitating lexical access as its canonical [p] counterpart. This runs contrary to an account of variant recognition that is driven purely by frequency of occurrence of variants in contexts (e.g., exemplar models: Goldinger, 1996). Sumner and Samuel (2005) also found that there was no advantage for [ʔ͡t] in terms of lexical access, despite it being more frequent in coda contexts. In our study, we similarly did not find any facilitation of word recognition when listeners heard glottalized [ʔ͡t] compared to non-glottalized [t]. These results are consistent with the robust finding that canonical forms are privileged, despite their actual production frequency across contexts (McLennan et al., 2003, 2005; Ranbom & Connine, 2007; Pitt, 2009; Pitt et al., 2011; Ranbom et al., 2009; Tucker, 2011).
Finally, our findings provide evidence that glottalization is not associated with voiced stops, as evidenced by slower and poorer recognition of target words ending in coda [ʔ͡d] and [ʔ͡b]. That is, glottalized voiced stops are less appropriate variants of /d/ and /b/. It is an open question, however, how much of a perceptual mismatch glottalized voiced stops are. Previous findings from word recognition studies have shown that the degree to which word recognition is impeded in both adult and child listeners is dependent on the degree of mismatch between an auditory signal and the lexical representation of a word (e.g., Andruski et al., 1994; McMurray et al., 2002; Gaskell, 2003; White & Morgan, 2008), with smaller mismatches resulting in a smaller decrement in performance. Given that glottalization is not phonemic in English, it is possible that glottalized voiced stops are a smaller perceptual mismatch for stored representations of words with coda voiced stops than whole phoneme substitutions (e.g., mog for mob). Future work will aim to address this possibility.
4.4. Conclusion and future directions
Using a printed text version of the visual world paradigm, we investigated the perceptual consequences of coda glottalization in American English. Previous researchers have claimed that glottalized coda stops arise from the speaker’s need to enhance voicelessness (or /t/ in particular) in environments where these stops are likely to become voiced or to overlap with other sounds (Pierrehumbert, 1994, 1995; Keyser & Stevens, 2006; Stevens & Keyser, 2010). Listeners treat glottalized variants of /p/ and /t/ as equally good instances of these sounds, and they treat glottalized variants of /b/ and /d/ either as unexpected variants or as mispronunciations. However, listeners do not use glottalization to fixate more on /t/ or /p/, which implies that glottalization is not used to enhance the acoustic attributes of these sounds and consequently facilitate word recognition. More generally, these results show how the visual world eye-tracking paradigm can be used to examine assumptions of explanatory models of phonological variation that predict facilitative or inhibitory effects on word recognition.
The additional file for this article can be found as follows:
Experimental stimuli list: Test items. DOI: https://doi.org/10.5334/labphon.70.s1
- , where N is the total number of samples in a given time bin and Y is the number of samples that fall in a given target interest area (Barr, 2008). [^]
- Although Oleson et al. (2015) present a by-subject analysis, we chose here to conduct a by-item analysis for reasons of statistical power. Given that each subject contributes few data points to each experimental condition, the bdots package fails to adequately model the by-subject data. Thus, our by-items analysis allows us to include more data points per item, allowing the bdots package to more successfully fit the fixation curve. [^]
We are grateful to Associate Editor Holger Mitterer and two anonymous reviewers for their helpful feedback, and to Kip Wilson for editorial assistance. We would also like to thank Amanda Ritchart, who recorded the stimuli, Scott Ragland and Stefan Wooding for their help with the stimuli resynthesis, as well as members of the UCLA Phonetics Lab and Sarah Creel for comments and discussion. An earlier version of this project was presented at the Fall 2016 meeting of the Acoustical Society of America.
The authors have no competing interests to declare.
Allopenna, P. D., Magnuson, J. S., & Tanenhaus, M. K. 1998. Tracking the time course of spoken word recognition using eye movements: Evidence for continuous mapping models. Journal of Memory and Language, 38, 419–439. DOI: http://doi.org/10.1006/jmla.1997.2558
Andruski, J. E., Blumstein, S. E., & Burton, M. 1994. The effect of subphonetic differences on lexical access. Cognition, 52, 163–187. DOI: http://doi.org/10.1016/0010-0277(94)90042-6
Barr, D. J. 2008. Analyzing ‘visual world’ eyetracking data using multilevel logistic regression. Journal of Memory and Language, 59, 457–474. DOI: http://doi.org/10.1016/j.jml.2007.09.002
Bates, D., Maechler, M., Bolker, B., & Walker, S. 2015. Fitting Linear Mixed-Effects Models Using lme4. Journal of Statistical Software, 67, 1–48. DOI: http://doi.org/10.18637/jss.v067.i01
Beddor, P. S. 1993. The perception of nasal vowels. In: Huffman, M. K., & Krakow, R. A. (eds.). Nasals, Nasalization, and the Velum, 171–196. Academic Press. DOI: http://doi.org/10.1016/B978-0-12-360380-7.50011-9
Beddor, P. S., McGowan, K. B., Boland, J. E., Coetzee, A. W., & Brasher, A. 2013. The time course of perception of coarticulation. Journal of the Acoustical Society of America, 133, 2350–2366. DOI: http://doi.org/10.1121/1.4794366
Boersma, P., & Weenink, D. 2015. Praat: Doing phonetics by computer [Computer program]. Version 5.4.04. Retrieved January 5, 2015 from: http://www.praat.org/.
Brouwer, S., Mitterer, H., & Huettig, F. 2012. Can hearing puter activate pupil? Phonological competition and the processing of reduced spoken words in spontaneous conversations. The Quarterly Journal of Experimental Psychology, 65, 2193–2220. DOI: http://doi.org/10.1080/17470218.2012.693109
Brysbaert, M., & New, B. 2009. Moving beyond Kučera and Francis: A critical evaluation of current word frequency norms and the introduction of a new and improved word frequency measure for American English. Behavior Research Methods, 41, 977–990. DOI: http://doi.org/10.3758/BRM.41.4.977
Cohen Priva, U. 2017. Informativity and the actuation of lenition. Language, 93, 569–597. DOI: http://doi.org/10.1353/lan.2017.0037
Dahan, D., Magnuson, J. S., Tanenhaus, M. K., & Hogan, E. M. 2001. Subcategorical mismatches and the time course of lexical access: Evidence for lexical competition. Language and Cognitive Processes, 16, 507–534. DOI: http://doi.org/10.1080/01690960143000074
Dahan, D., & Tanenhaus, M. K. 2004. Continuous mapping from sound to meaning in spoken-language comprehension: Immediate effects of verb-based thematic constraints. Journal of Experimental Psychology: Learning, Memory, and Cognition, 30, 498–513. DOI: http://doi.org/10.1037/0278-7322.214.171.1248
Davidson, L., & Erker, D. 2014. Hiatus resolution in American English: The case against glide insertion. Language, 90, 482–514. DOI: http://doi.org/10.1353/lan.2014.0028
Dilley, L., Shattuck-Hufnagel, S., & Ostendorf, M. 1996. Glottalization of word-initial vowels as a function of prosodic structure. Journal of Phonetics, 24, 423–444. DOI: http://doi.org/10.1006/jpho.1996.0023
Esling, J. H., Fraser, K. E., & Harris, J. G. 2005. Glottal stop, glottalized resonants, and pharyngeals: A reinterpretation with evidence from a laryngoscopic study of Nuuchahnulth (Nootka). Journal of Phonetics, 33, 383–410. DOI: http://doi.org/10.1016/j.wocn.2005.01.003
Esposito, C. M. 2012. An acoustic and electroglottographic study of White Hmong phonation. Journal of Phonetics, 40, 466–476. DOI: http://doi.org/10.1016/j.wocn.2012.02.007
Foulkes, P., & Docherty, G. 2007. Phonological variation in England. In: Britain, D. (ed.). Language in the British Isles, 52–74. Cambridge: Cambridge University Press. DOI: http://doi.org/10.1017/CBO9780511620782.005
Garellek, M. 2012. The timing and sequencing of coarticulated non-modal phonation in English and White Hmong. Journal of Phonetics, 40, 152–161. DOI: http://doi.org/10.1016/j.wocn.2011.10.003
Garellek, M. 2013. Production and perception of glottal stops (Doctoral dissertation). Los Angeles, CA: UCLA.
Garellek, M. 2015. Perception of glottalization and phrase-final creak. Journal of the Acoustical Society of America, 137, 822–831. DOI: http://doi.org/10.1121/1.4906155
Garellek, M. (to appear). The phonetics of voice. In: Katz, W., & Assmann, P. (eds.), Handbook of Phonetics. New York: Routledge.
Garellek, M., Keating, P., Esposito, C. M., & Kreiman, J. 2013. Voice quality and tone identification in White Hmong. Journal of the Acoustical Society of America, 133, 1078–1089. DOI: http://doi.org/10.1121/1.4773259
Garellek, M., & Seyfarth, S. 2016. Acoustic differences between English /t/ glottalization and phrasal creak. In: Proceedings of Interspeech 2016, 1054–1058. San Francisco. DOI: http://doi.org/10.21437/Interspeech.2016-1472
Gaskell, M. G. 2003. Modelling regressive and progressive effects of assimilation in speech perception. Journal of Phonetics, 31, 447–463. DOI: http://doi.org/10.1016/S0095-4470(03)00012-3
Gerratt, B. R., & Kreiman, J. 2001. Toward a taxonomy of nonmodal phonation. Journal of Phonetics, 29, 365–381. DOI: http://doi.org/10.1006/jpho.2001.0149
Goldinger, S. D. 1996. Words and voices: Episodic traces in spoken word identification and recognition memory. Journal of experimental psychology: Learning, Memory, and Cognition, 22, 1166–1183. DOI: http://doi.org/10.1037/0278-73126.96.36.1996
Gordeeva, O. B., & Scobbie, J. M. 2013. A phonetically versatile contrast: Pulmonic and glottalic voicelessness in Scottish English obstruents and voice quality. Journal of the International Phonetic Association, 43, 249–271. DOI: http://doi.org/10.1017/S0025100313000200
Gow, D. W., & McMurray, B. 2007. Word recognition and phonology: The case of English coronal place assimilation. Papers in laboratory phonology, 9, 173–200.
Huffman, M. K. 2005. Segmental and prosodic effects on coda glottalization. Journal of Phonetics, 33, 335–362. DOI: http://doi.org/10.1016/j.wocn.2005.02.004
Jongman, A., Blumstein, S. E., & Lahiri, A. 1985. Acoustic properties for dental and alveolar stop consonants: A cross-language study. Journal of Phonetics, 13, 235–251.
Keating, P., Garellek, M., & Kreiman, J. 2015. Acoustic properties of different kinds of creaky voice. In: Proceedings of the 18th International Congress of Phonetic Sciences. Glasgow.
Keyser, S. J., & Stevens, K. N. 2006. Enhancement and overlap in the speech chain. Language, 82, 33–63. DOI: http://doi.org/10.1353/lan.2006.0051
Kuznetsova, A., Brockhoff, P. B., & Christensen, R. H. B. 2016. lmerTest: Tests in Linear Mixed Effects Models. R package version 2.0-30. https://CRAN.R-project.org/package=lmerTest.
Luce, P. A., & Charles-Luce, J. 1985. Contextual effects on vowel duration, closure duration, and the consonant/vowel ratio in speech production. Journal of the Acoustical Society of America, 78, 1949–1957. DOI: http://doi.org/10.1121/1.392651
Matin, E., Shao, K. C., & Boff, K. R. 1993. Saccadic overhead: Information-processing time with and without saccades. Perception & Psychophysics, 53, 372–380. DOI: http://doi.org/10.3758/BF03206780
McLennan, C. T., Luce, P. A., & Charles-Luce, J. 2003. Representation of lexical form. Journal of Experimental Psychology, Learning, Memory, and Cognition, 29, 539–553. DOI: http://doi.org/10.1037/0278-73188.8.131.529
McLennan, C. T., Luce, P. A., & Charles-Luce, J. 2005. Representation of lexical form: Evidence from studies of sublexical ambiguity. Journal of Experimental Psychology: Human Perception and Performance, 31, 1308–1314. DOI: http://doi.org/10.1037/0096-15184.108.40.2068
McMurray, B., Tanenhaus, M. K., & Aslin, R. N. 2002. Gradient effects of within-category phonetic variation on lexical access. Cognition, 86, B33–B42. DOI: http://doi.org/10.1016/S0010-0277(02)00157-9
McQueen, J. M., & Viebahn, M. C. 2007. Tracking recognition of spoken words by tracking looks to printed words. The Quarterly Journal of Experimental Psychology, 60, 661–671. DOI: http://doi.org/10.1080/17470210601183890
Mitterer, H. 2011. The mental lexicon is fully specified: Evidence from eye-tracking. Journal of Experimental Psychology: Human Perception and Performance, 37, 496–513. DOI: http://doi.org/10.1037/a0020989
Oleson, J. J., Cavanaugh, J. E., McMurray, B., & Brown, G. 2015. Detecting time-specific differences between temporal nonlinear curves: Analyzing data from the visual world paradigm. Statistical Methods in Medical Research, 1–22.
Penney, J., Cox, F., Miles, K., & Palethorpe, S. 2018. Glottalisation as a cue to coda consonant voicing in Australian English. Journal of Phonetics, 66, 161–184. DOI: http://doi.org/10.1016/j.wocn.2017.10.001
Peterson, G. E., & Lehiste, I. 1960. Duration of syllable nuclei in English. Journal of the Acoustical Society of America, 32, 693–703. DOI: http://doi.org/10.1121/1.1908183
Pierrehumbert, J. 1994. Knowledge of variation. In: Papers from the parasession on variation, 30th meeting of the Chicago Linguistic Society, 232–256. Chicago: Chicago Linguistic Society.
Pierrehumbert, J. 1995. Prosodic effects on glottal allophones. In: Fujimura, O., & Hirano, M. (eds.), Vocal fold physiology: Voice quality control, 39–60. San Diego: Singular Publishing Group.
Pitt, M. A. 2009. How are pronunciation variants of spoken words recognized? A test of generalization to newly learned words. Journal of Memory and Language, 61, 19–36. DOI: http://doi.org/10.1016/j.jml.2009.02.005
Pitt, M. A., Dilley, L., Johnson, K., Kiesling, S., Raymond, W., Hume, E., & Fosler-Lussier, E. 2007. Buckeye Corpus of Conversational Speech (2nd release). Department of Psychology, Ohio State University Columbus, OH.
Pitt, M. A., Dilley, L., & Tat, M. 2011. Exploring the role of exposure frequency in recognizing pronunciation variants. Journal of Phonetics, 39, 304–311. DOI: http://doi.org/10.1016/j.wocn.2010.07.004
Ranbom, L. J., & Connine, C. M. 2007. Lexical representation of phonological variation in spoken word recognition. Journal of Memory and Language, 57, 273–298. DOI: http://doi.org/10.1016/j.jml.2007.04.001
Ranbom, L. J., Connine, C. M., & Yudman, E. M. 2009. Is phonological context always used to recognize variant forms in spoken word recognition? The role of variant frequency and context distribution. Journal of Experimental Psychology: Human Perception and Performance, 35, 1205–1220. DOI: http://doi.org/10.1037/a0015022
R Core Team. 2015. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing.
Redi, L., & Shattuck-Hufnagel, S. 2001. Variation in the realization of glottalization in normal speakers. Journal of Phonetics, 29, 407–429. DOI: http://doi.org/10.1006/jpho.2001.0145
Roach, P. J. 1973. Glottalization of English /p/, /t/, /k/ and /tS/ – a re-examination. Journal of the International Phonetic Association, 3, 10–21. DOI: http://doi.org/10.1017/S0025100300000633
Salverda, A. P., Kleinschmidt, D., & Tanenhaus, M. K. 2014. Immediate effects of anticipatory coarticulation in spoken-word recognition. Journal of Memory and Language, 71, 145–163. DOI: http://doi.org/10.1016/j.jml.2013.11.002
Seedorff, M., Oleson, J. J., Brown, G., Cavanaugh, J., & McMurray, B. 2017. bdots: Bootstrapped differences of time series. R package version 0.1.13. https://CRAN.R-project.org/package=bdots.
Seyfarth, S., & Garellek, M. 2015. Coda glottalization in American English. In: Proceedings of the 18th International Congress of Phonetic Sciences. Glasgow.
Stevens, K. N., & Keyser, S. J. 1989. Primary features and their enhancement in consonants. Language, 65, 81–106. DOI: http://doi.org/10.2307/414843
Stevens, K. N., & Keyser, S. J. 2010. Quantal theory, enhancement and overlap. Journal of Phonetics, 38, 10–19. DOI: http://doi.org/10.1016/j.wocn.2008.10.004
Styler, W. 2017. On the acoustical features of vowel nasality in English and French. Journal of the Acoustical Society of America, 142, 2469–2482. DOI: http://doi.org/10.1121/1.5008854
Sumner, M., & Samuel, A. G. 2005. Perception and representation of regular variation: The case of final /t/. Journal of Memory and Language, 52, 322–338. DOI: http://doi.org/10.1016/j.jml.2004.11.004
Tucker, B. V. 2011. The effect of reduction on the processing of flaps and /g/ in isolated words. Journal of Phonetics, 39, 312–318. DOI: http://doi.org/10.1016/j.wocn.2010.12.001
Westbury, J. R., & Keating, P. A. 1986. On the naturalness of stop consonant voicing. Journal of Linguistics, 22, 145–166. DOI: http://doi.org/10.1017/S0022226700010598
White, K. S., & Morgan, J. L. 2008. Sub-segmental detail in early lexical representations. Journal of Memory and Language, 59, 114–132. DOI: http://doi.org/10.1016/j.jml.2008.03.001
Zellou, G. 2017. Individual differences in the production of nasal coarticulation and perceptual compensation. Journal of Phonetics, 61, 13–29. DOI: http://doi.org/10.1016/j.wocn.2016.12.002