Two /b/ or not “too bee”: Gradient sensitivity to subphonemic variation, categorical perception and the effect of task. Bob McMurray Michael K. Tanenhaus.

Slides:



Advertisements
Similar presentations
Accessing spoken words: the importance of word onsets
Advertisements

All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
Marslen-Wilson Big Question: “What processes take place during the period that the sensory information is accumulating for the listener” during spoken.
Human Speech Recognition Julia Hirschberg CS4706 (thanks to John-Paul Hosum for some slides)
CS 551/651: Structure of Spoken Language Lecture 12: Tests of Human Speech Perception John-Paul Hosom Fall 2008.
Eye Movements and Spoken Language Comprehension: effects of visual context on syntactic ambiguity resolution Spivey et al. (2002) Psych 526 Eun-Kyung Lee.
Infant sensitivity to distributional information can affect phonetic discrimination Jessica Maye, Janet F. Werker, LouAnn Gerken A brief article from Cognition.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Prosodic facilitation and interference in the resolution of temporary syntactic closure ambiguity Kjelgaard & Speer 1999 Kent Lee Ψ 526b 16 March 2006.
Language Comprehension Speech Perception Semantic Processing & Naming Deficits.
Spoken Word Recognition 1 Language Use and Understanding.
SPEECH PERCEPTION 2 DAY 17 – OCT 4, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
The Neuroscience of Language. What is language? What is it for? Rapid efficient communication – (as such, other kinds of communication might be called.
Using prosody to avoid ambiguity: Effects of speaker awareness and referential context Snedeker and Trueswell (2003) Psych 526 Eun-Kyung Lee.
And a big thanks to Julie Markant
Speech perception 2 Perceptual organization of speech.
9/22/10Psyc / Ling / Comm 525 Fall 10 Semantic Priming (Phenomenon & Tool)...armkitchentree Related prime >doctoractor < Unrelated prime nurse floor...
Splice: From vowel offset to vowel onset FIG 3. Example of stimulus spliced from the repetitive syllables. EXPERIMENT 2 (Voicing ID) METHOD Speech materials:
Development of Speech Perception. Issues in the development of speech perception Are the mechanisms peculiar to speech perception evident in young infants?
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Introduction Complex words may be either (a) stored as full forms in the mental lexicon, or (b) undergo decomposition into their constituent morphemes.
Auditory Word Recognition
Experiment 2: MEG Study Materials and Methods: 11 right-handed subjects with 20:20 vision were run. 3 subjects’ data was discarded because of poor performance.
How General is Lexically-Driven Perceptual Learning of Phonetic Identity? Tanya Kraljic and Arthur G. Samuel Our Questions (e.g., learning a particular.
Exam 1 Monday, Tuesday, Wednesday next week WebCT testing centre Covers everything up to and including hearing (i.e. this lecture)
Cognitive Processes PSY 334 Chapter 2 – Perception April 9, 2003.
TEMPLATE DESIGN © Listener’s variation in phoneme category boundary as a source of sound change: a case of /u/-fronting.
PSY 369: Psycholinguistics
Interrupted speech perception Su-Hyun Jin, Ph.D. University of Texas & Peggy B. Nelson, Ph.D. University of Minnesota.
Sound and Speech. The vocal tract Figures from Graddol et al.
Influence of Word Class Proportion on Cerebral Asymmetries for High and Low Imagery Words Christine Chiarello 1, Connie Shears 2, Stella Liu 3, and Natalie.
Language Comprehension Speech Perception Naming Deficits.
The Perception of Speech
Language Comprehension Speech Perception Meaning Representation.
Cognitive Processes PSY 334 Chapter 2 – Perception.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
SPOKEN LANGUAGE COMPREHENSION Anne Cutler Addendum: How to study issues in spoken language comprehension.
Sebastián-Gallés, N. & Bosch, L. (2009) Developmental shift in the discrimination of vowel contrasts in bilingual infants: is the distributional account.
Experiments concerning boundary tone perception in German 3 rd Workshop of the SPP-1234 Potsdam, 7 th January 2009 Presentation of the Stuttgart Project.
Speech Perception 4/6/00 Acoustic-Perceptual Invariance in Speech Perceptual Constancy or Perceptual Invariance: –Perpetual constancy is necessary, however,
Statistical learning, cross- constraints, and the acquisition of speech categories: a computational approach. Joseph Toscano & Bob McMurray Psychology.
1 Speech Perception 3/30/00. 2 Speech Perception How do we perceive speech? –Multifaceted process –Not fully understood –Models & theories attempt to.
Speech Perception1 Fricatives and Affricates We will be looking at acoustic cues in terms of … –Manner –Place –voicing.
Adaptive Design of Speech Sound Systems Randy Diehl In collaboration with Bjőrn Lindblom, Carl Creeger, Lori Holt, and Andrew Lotto.
Speech Science Fall 2009 Oct 28, Outline Acoustical characteristics of Nasal Speech Sounds Stop Consonants Fricatives Affricates.
SPEECH PERCEPTION DAY 16 – OCT 2, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
LATERALIZATION OF PHONOLOGY 2 DAY 23 – OCT 21, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
Epenthetic vowels in Japanese: a perceptual illusion? Emmanual Dupoux, et al (1999) By Carl O’Toole.
Pragmatically-guided perceptual learning Tanya Kraljic, Arty Samuel, Susan Brennan Adaptation Project mini-Conference, May 7, 2007.
From Sound to Sense and back again: The integration of lexical and speech processes From Sound to Sense and back again: The integration of lexical and.
The long-term retention of fine- grained phonetic details: evidence from a second language voice identification training task Steve Winters CAA Presentation.
The New Normal: Goodness Judgments of Non-Invariant Speech Julia Drouin, Speech, Language and Hearing Sciences & Psychology, Dr.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
Neurophysiologic correlates of cross-language phonetic perception LING 7912 Professor Nina Kazanina.
The effects of working memory load on negative priming in an N-back task Ewald Neumann Brain-Inspired Cognitive Systems (BICS) July, 2010.
Stop-Consonant Perception in 7.5-month-olds: Evidence for gradient categories Bob McMurray & Richard N. Aslin Department of Brain and Cognitive Sciences.
AUDITORY CORTEX 4 SEPT 21, 2015 – DAY 12 Brain & Language LING NSCI Fall 2015.
Experimental Control Definition Is a predictable change in behavior (dependent variable) that can be reliably produced by the systematic manipulation.
WebCT You will find a link to WebCT under the “Current Students” heading on It is your responsibility to know how to work WebCT!
Nuclear Accent Shape and the Perception of Syllable Pitch Rachael-Anne Knight LAGB 16 April 2003.
AUDITORY CORTEX 1 SEPT 11, 2015 – DAY 8 Brain & Language LING NSCI Fall 2015.
Against formal phonology (Port and Leary).  Generative phonology assumes:  Units (phones) are discrete (not continuous, not variable)  Phonetic space.
Chapter 11 Language. Some Questions to Consider How do we understand individual words, and how are words combined to create sentences? How can we understand.
Effects of Mispronunciation on Spoken Word Recognition in
S. Kramer1, K. Tucker1, A.L. Moro1, E. Service1, J.F. Connolly1
Phonological Priming and Lexical Access in Spoken Word Recognition
Cognitive Processes PSY 334
The lexical/phonetic interface: Evidence for gradient effects of within-category VOT on lexical access Bob McMurray Richard N. Aslin Mickey K. TanenMouse.
Phonological Priming and Lexical Access in Spoken Word Recognition
Speech Perception (acoustic cues)
Presentation transcript:

Two /b/ or not “too bee”: Gradient sensitivity to subphonemic variation, categorical perception and the effect of task. Bob McMurray Michael K. Tanenhaus Michael J. Spivey Richard N. Aslin Dana Subik With thanks to

Invariance, Covariance and Gradient Sensitivity in speech perception. Categorical Perception and other previous research. Experiment 1: Gradient sensitivity in Word Recognition Experiment 2-5: The effect of experimental task Targets & competitors, gradient sensitivity and temporal dynamics, Conclusions Outline

Problem of Invariance Phonetic features are correlated with many acoustic realizations. Acoustic realization of a phonetic feature depends on context. How do we extract invariant linguistic representations from a variable acoustic signal? What properties of the signal provide an invariant mapping to linguistic representations? How do we extract discrete units from a graded signal?

Problem of Invariance Two Solutions Motor Theory: acoustic invariance does not exist, but specialized mechanisms allow us to unpack speech into invariant motor representations (Liberman & Mattingly, 1985; Fowler, 1986). Acoustic Invariance: better computational methods and neurologically inspired models may find invariant acoustic properties of the signal (Blumstein,1998; Sussman et al, 1998)

The Fundamental Approach How do we pay attention to the right parts of the signal and ignore the variation? However, recent work suggests that this the variation is actually highly informative covariation. Rethinking Invariance

In measurements of productions, effects of speaking rate on VOT (e.g. Kessinger & Blumstein) prosodic domain and VOT and articulatory strength (Fougeron and Keating) Place of articulation and vowel quality 5 syllables away (Local) Between-consonant coarticulation (Mann & Repp) suggest that a system sensitive to fine grained detail could take advantage of all of this information. Rethinking Invariance

Speech perception shows probabilistic effects of many information sources: Lexical Context Spectral vs. Temporal Cues Visual Information Transition Statistics Speech RateStimulus Naturalness Sentential ContextCompensatory Coarticulation EmbeddingsSyllabic Stress Lexical StressPhrasal Stress A system that was sensitive to fine-grained acoustic detail might be much more efficient than one that was not. Tracking covariance may help solve the problem of invariance. Rethinking Invariance

What sort of sensitivity is needed? Gradient Sensitivity: As fundamentally graded acoustic information changes (even changes that still result in the same “category”), activation of lexical or sublexical representation changes monotonically. Activation of linguistic units reflects the probability that a that unit is instantiated by the acoustic signal.

Categorical Perception CP suggests listeners do not show gradient sensitivity to subphonemic information. Sharp identification of speech sounds on a continuum ID (%/pa/) VOT PB % /p/ B P Discrimination Discrimination poor within a phonetic category

Evidence for Categorical Perception Supported by: Work on VOT and place of articulation. Ubiquity of steep identification functions. Recent electrophysiological data (e.g. Philips, Pellathy, Marantz, Yellin, Wexler, Poeppel, McGinnis & Roberts, 2000; Sharma & Dorman, 1999)

Revisiting Categorical Perception? Evidence against CP comes from Discrimination Tasks Pisoni and Tash (1974) Pisoni & Lazarus (1974) Carney, Widin & Viemeister (1977) Training Samuel (1977) Pisoni, Aslin, Perey & Hennessy (1982) Goodness Ratings Miller (1997)Massaro & Cohen, 1983 Only goodness ratings show any hint of gradiency. No gradient effects from identification tasks. But, 2AFC metalinguistic tasks may underestimate sensitivity to subphonemic acoustic information

Lexical sensitivity Andruski, Blumstein & Burton (1994) Created stimuli that were either voiceless, 1/3 or 2/3 voiced. 2/3 voiced stimuli primed semantic associates more weakly than fully voiceless or 1/3 voiced tokens First demonstration of lexical sensitivity to natural variation in consonants. However: 2/3 voiced stimuli were close to category boundary. No evidence for gradiency—difference between 2 items. Hard to interpret temporal dynamics in priming tasks.

Remaining Questions Is sensitivity to subphonemic differences gradient? Is it symmetrical (I.e. gradiency on both sides of category boundary)? Are differences preserved long enough to be usefully combined with subsequent input? Perhaps a more sensitive measure….

250 Hz realtime stream of eye positions. Parsed into Saccades, Fixations, Blinks, etc… Head movement compensation. Output in ~screen coordinates. Head-Tracker Cam Monitor IR Headtracker Emitters Eyetracker Computer Subject Computer Computers connected via Ethernet 2 Eye cameras HeadEye-TrackingEye-Tracking

Eye-TrackingEye-Tracking Fixations to object in response to spoken instructions: are time locked to incoming information (Tanenhaus, Spivey-Knowlton, Ebehart and Sedivy, 1995) can be easily mapped onto lexical activation from models like TRACE (Allopenna, Magnuson and Tanenhaus, 1998) show effects of non-displayed competitors (Dahan, Magnuson, Tanenhaus & Hogen) provide a glimpse at how activation for competitors unfolds in parallel over time.

Experiment 1 Lexical Identification “too bee” Experiment 1 Lexical Identification “too bee” Can we use eye-tracking methodologies to find evidence for graded perception of VOT?

Experiment 1: Lexical Identification Six 9-step /ba/ - /pa/ VOT continuum (0-40ms) Bear/PearBeach/Peach Butter/PutterBale/Pale Bump/PumpBomb/Palm 12 L- and Sh- Filler items LeafLampLadderLock LipLegSharkShip ShirtShoeShellSheep Identification indicated by mouse click on picture Eye movements monitored at 250 hz 17 Subjects

A moment to view the items Experiment 1: Lexical Identification

500 ms later Experiment 1: Lexical Identification

Bear

Experiment 1: Identification Results By subject: /- 1.33ms By item: /- 1.24ms VOT (ms) proportion /p/ BP High agreement across subjects and items for category boundary

Analysis of fixations VOT (ms) proportion /p/ BP Actual Exp2 Data Trials with low- frequency response excluded. ID Function after filtering Yields a “perfect” categorization function.

+ Target = bug Competitor = bus Unrelated = cat, fish Time 200 ms Trials Analysis of fixations

Experiment 1: Eye Movement Results Time (ms) More looks to competitor than unrelated items VOT=0 Response= VOT=40 Response= Fixation proportion

e.g. Given that the subject heard bomb clicked on “bomb”… time Fixation proportion target competitor How often was the Subject looking at the “palm”? time Fixation proportion target competitor Categorical Results Gradient Effect Analysis of fixations Gradient “competitor” effects

ms 5 ms 10 ms 15 ms VOT ms 25 ms 30 ms 35 ms 40 ms VOT Fixation proportion Time since word onset (ms) Experiment 1: Eye Movement Results Smaller effect on the amplitude of activation—more effect on the duration: Competitors stay active longer as VOT approaches the category boundary. Response= Gradient competitor effects of VOT?

Experiment 1: Gradiency? VOT (ms) Fixation proportion Looks to Gradient Sensitivity “Categorical” Perception Andruski et al (schematic)

Experiment 1: Eye Movement Results VOT (ms) Fixation proportion Category Boundary B: p=.017* P: p<.0001*** Clear effects of VOT Response= Looks to Linear TrendB: p=.023*P: p=.002**

Experiment 1: Eye Movement Results VOT (ms) Fixation proportion Category Boundary Response= Looks to B: p=.017* P: p<.0001*** Clear effects of VOT Unambiguous Stimuli Only Linear TrendB: p=.023*P: p=.002**

Experiment 1: Results and Conclusions Subphonemic acoustic differences in VOT affect lexical activation. Gradient effect of VOT on looks to the competitor. Effect seems to be long-lasting (we’ll get back to that). Effect holds even for unambiguous stimuli. Conservative Test Filter out “incorrect” responses. Use unambiguous stimuli only.

Why was it so hard to find evidence for gradiency in CP tasks? However… * Steep identification function consistently replicated. What aspects of the task affect our ability to see gradient sensitivity? Phoneme ID vs. Lexical ID? Number of Alternatives? Type of Stimuli? Sensitivity of response measure

Experiment 2 Categorical Perception 2 /b/, not “too bee” Experiment 2 Categorical Perception 2 /b/, not “too bee” What can the eye-tracking paradigm reveal about ordinary phoneme identification experiments?

Replicates “classic” task: 9-step /ba/ - /pa/ VOT continuum (0-40ms) 2AFC Identification indicated by mouse click. Eye movements monitored at 250 hz. 17 Subjects Experiment 2: Categorical Perception

12 BP Ba 3 Experiment 2: Categorical Perception

Experiment 2: Identification Results VOT (ms) proportion /p/ Exp 2: BP Exp 1: Words Phoneme ID function steeper BP:17.5 +/-.83ms Words subject : /-1.33ms Words item : /- 1.24ms Boundaries Category boundaries are the same. BP

VOT (ms) Proportion of /p/ response Experiment 2: Data Analysis Trials with low-frequency response excluded. Effectively yields a “perfect” categorization function. Actual ID Function Effective ID Function

Time (ms) ms 5 ms 10 ms 15 ms VOT 20 ms 25 ms VOT 30 ms 35 ms 40 ms Some hints of gradiency for /p/. Even less for /b/. Difference between stimuli near boundary and endpoints. Perhaps more for /p/. Experiment 2: Eye movement data Fixation proportion Response = B Response = P Looks to P Looks to B

Experiment 2: Eye movement data /b/:p =.044* p trend =.055 /p/:p<.001***p trend =.005*** Could be driven by differences near category boundary VOT (ms) Category Boundary Response=B Looks to P Response=B Looks to P Fixation proportion

Experiment 2: Eye movement data Unambiguous Stimuli Only /b/:p =.884p trend =.678 /p/:p =.013*p trend =.003*** VOT (ms) Category Boundary Response=B Looks to P Response=B Looks to P Fixation proportion

Experiment 2: Results and Conclusions Very steep slope for mouse response curves. consistent with traditional results Identical category boundary to experiment 1 validates stimuli Small difference between stimuli near category boundary and others. similar to Pisoni & Tash, Andruski, et al. Gradient effect weak for /ba/, moderate for /pa/

Experiment 3 Number of Response Alternatives Not 2 but /b/? Experiment 3 Number of Response Alternatives Not 2 but /b/? compare to experiment 2 (BaPa)

Experiment 3: BaPaLaSha Given the strong evidence for gradiency in Experiment 1 and the weaker evidence in Experiment 2, what is the effect of number of response alternatives? Same 9-step /ba/ - /pa/ VOT continuum (0-40ms) as experiment 2. La and Sha filler items added. 4AFC Identification indicated by mouse click. Button locations randomized between subjects. Eye movements monitored at 250 hz. 17 Subjects

Experiment 3: BaPaLaSha P BSh L La

VOT proportion /p/ response Experiment 3: Identification Results Exp. 2 (BaPa) Exp. 1 (words) Exp. 3 (BaPaLaSha) Number of response alternatives accounts for some of the difference in slope.

Effective ID Function VOT (ms) Proportion of /p/ response Experiment 3: Data Analysis Trials with low-frequency response excluded. Effectively yields a “perfect” categorization function. Actual ID Function

More looks to competitor than unrelated stimuli (p<.001). Eye movements in “phoneme ID” tasks are sensitive to acoustic similarity. Experiment 3: Eye movement data Time (ms) VOT=0 Response=b VOT=40 Response=P Fixation proportion B P UR

Time (ms) 0 ms 5 ms 10 ms 15 ms VOT Difference between stimuli near boundary and endpoints Experiment 3: Eye movement data Fixation proportion ms 25 ms VOT 30 ms 35 ms 40 ms Response = B Response = P Looks to P Looks to B

Experiment 3: Eye movement data Close but no star: Nothing reaches significance /b/:p=.055 p trend =.068 /p/:p=.510 p trend = VOT (ms) Category Boundary Response=B Looks to P Response=B Looks to P Fixation proportion

Experiment 3: Eye movement data Unambiguous Stimuli Only: even worse /b/:p=.374 p trend =.419 /p/:p=.356 p trend = VOT (ms) Category Boundary Response=B Looks to P Response=B Looks to P Fixation proportion

Eye movements in phoneme ID tasks are sensitive to acoustic similarity between target and competitor. Number of alternatives explains some of differences in ID function. VERY weak subphonemic effects on lexical activation. Experiment 3: Results

Experiment 4 Response Type “too” /b/ Experiment 4 Response Type “too” /b/ Is there a difference between phoneme and lexical identification tasks? compare to experiment 1 (words)

Experiment 4: Response Type Same 6 VOT continua (0-40ms) as experiment 1 beach/peachbear/pearbomb/palm bale/pailbump/pumpbutter/putter Same 12 L- and Sh- filler items. 4AFC phoneme identification indicated by mouse click. Button locations randomized between subjects. Eye movements monitored at 250 hz. 17 Subjects

Experiment 4: Response Type P BSh L Ship

VOT proportion /p/ response Experiment 4: Identification Results Exp. 2 (BaPa) Exp. 1 (words) Exp. 4 (Response Type) Similar category boundary and slope to Exp 1 Exp 1: /- 1.33ms Exp 4: /- 1.52ms

ms 5 ms 10 ms 15 ms VOT 20 ms 25 ms VOT 30 ms 35 ms 40 ms Time (ms) Fixation proportion Experiment 4: Eye movement data Small differences in the right direction Response = B Response = P Looks to P Looks to B

Experiment 4: Eye movement data Gradient effects using the whole range of stimuli /b/:p<.001 p trend =.002 /p/:p=.001 p trend = VOT (ms) Category Boundary Fixation proportion Response=B Looks to B Response=P Looks to B

Experiment 4: Eye movement data Marginal effects using “unambiguous” stimuli only. /b/:p=.074 p trend =.074 /p/:p=.137p trend = VOT (ms) Category Boundary Fixation proportion Response=B Looks to B Response=P Looks to B

Weaker subphonemic effect suggests that offline “metalinguistic” tasks are less sensitive to fine-grained phonetic detail than online tasks. Some detail is preserved in these tasks (at least with word stimuli)… Experiment 4: Results

Experiment 5 2AFC Words 2 “bee” Experiment 5 2AFC Words 2 “bee” Bringing it all together

Experiment 5: 2-Words Is the difference in ID curve slopes purely the result of number of response alternatives or does task play a role? Same 6 VOT continua (0-40ms) as experiment 1 beach/peachbear/pearbomb/palm bale/pailbump/pumpbutter/putter 0 filler items. 2AFC phoneme identification indicated by mouse click. Eye movements monitored at 250 hz.

Pear Experiment 5: Task

VOT proportion /p/ response Experiment 5: Identification Results Exp. 2 (BaPa) Exp. 1 (words) Exp. 5 (2-words) Similar category boundary and slope to Exp 1 Exp 1: /- 1.33ms Exp 5: /- 1.74ms BP

0 ms 5 ms 10 ms 15 ms VOT 20 ms 25 ms VOT 30 ms 35 ms 40 ms Time (ms) Fixation proportion Experiment 5: Eye movement data Response= Clean, but small, gradient effects for /p/ Effects for /b/ near the category boundary.

Experiment 5: Eye movement data Gradient effects using the whole range of stimuli /b/:p<.001 p trend =.005 /p/:p=.017 p trend = VOT (ms) Category Boundary Fixation proportion Response= Looks to Response= Looks to

Experiment 5: Eye movement data Weaker effects using the “prototypical” range /b/:p<.443 p trend =.802 /p/:p=.044* p trend = VOT (ms) Category Boundary Fixation proportion Response= Looks to Response= Looks to

Shallow ID curve slope suggests 2AFC alone is not enough to create steep slope: 2AFC- phoneme ID is needed. Weaker gradient effects: fixed response locations and no filler items make this task more “explicit”? Experiment 5: Results

Trying to make sense out of it all… Being and Nothingness? Trying to make sense out of it all… Being and Nothingness?

Slope of ID Function Exp 1 (words) Exp 2 (BaPa) Exp 3 (BaPaLaSha) Exp 4 Exp 5 (2 Words) Slope BP > BaPaLaSha > all others (p<.05) Words ~= Exp 4 ~= 2 Words (p>.1) 2AFC results in less sensitivity (in ID function) than 4AFC for non-word stimuli.

Gradient Effect across experiments 1 Words 2 BP 3 BPLS 4 Phoneme ID 5 2 words B ? P X B XX?X P XX All stimuli Without stimuli near c.b.

Pooled eye movement data VOT (ms) Fixation proportion Response=B Looks to P Response=P Looks to B Category Boundary Overall B: p vot.15 P: p vot.2

VOT (ms) Fixation proportion Response=B Looks to P Response=P Looks to B Category Boundary Without stimuli near category boundary B: p vot =.005p trend.1 P: p vot.2 Pooled eye movement data

Conclusions on Task Manipulations Phoneme ID tasks with non-words yield the sharpest categorization functions—may mask subphonemic sensitivity. Even within these tasks, the number of response alternatives makes a big difference. Identification Functions

“Natural”, 4AFC lexical identification provides cleanest evidence for gradiency (measured by fixations to the competitor) for both /p/ and /b/ halves of continuum. Conclusions on Task Manipulations Competitor Effects (eye-movements) All experiments offer evidence of subphonemic sensitivity when we include stimuli near the category boundary. Eye-movements provide much more sensitive measure for assessing the role of fine-grained phonetic detail.

Most experiments showed weak evidence for gradient effect, but larger effects for /p/ than /b/. Conclusions on Task Manipulations Competitor Effects (eye-movements) Differences in the variance of the distribution of /b/ and /p/ in the learning environment? (Lisker & Abramson, Gerken & Maye) Auditory locus? Double peaked firing in auditory cortex differs shows more VOT sensitivity to voiceless than voiced stops. (Steinshneider et al; Sharma et al) No one factor seems to account for presence or absence of gradient effect.

Targets and Competitors, Gradient Effects and Temporal Dynamics and a return to experiment 1

Targets and Competitors Why look at exclusively at the competitor? Do subphonemic differences affect activation of the target? Andruski et al suggests it does.

Experiment 1: Target Activation ms 5 ms 10 ms 15 ms VOT 20 ms 25 ms VOT 30 ms 35 ms 40 ms Response= Time (ms) Fixation proportion Target effects much weaker, even in Experiment 1 May be limited to range near category boundary.

Experiment 1: Target Activation VOT (ms) Category Boundary Fixation proportion Response= Looks to Response= Looks to Overall B: p vot =.035p trend =.103 P: p vot <.001p trend <.010

Experiment 1: Target Activation VOT (ms) Category Boundary Fixation proportion Response= Looks to Response= Looks to Unambiguous Stimuli Only B: p vot =.44p trend =.98 P: p vot =.33p trend =.22

Target Activation: Conclusions Target sensitivity to subphonemic differences is carried by differences between ambiguous and prototypical stimuli. Consistent with previous research Andruski et al: 2/3 voicing is close to ambiguous region (~27 ms). Pisoni & Tash: increased RT near boundary.

Target Activation: Conclusions Gradient sensitivity to subphonemic differences is stronger in competitor activation that target activation. Consistent with Misuirski, Blumstein, Rissman and Berman (in press) This makes sense: Degrading target activation isn’t likely to be helpful in word recognition. Augmenting competitor activation could be very helpful.

Gradiency and Time Phonetic context in speech perception isn’t simultaneous. Rate information (vowel length) arrives after consonant. Coarticulation occurs across multiple segments. Lexical information has a large scope than phonetic information. Simply tracking graded acoustic features is not enough. Graded activation of lexical or sublexical units must persist over time to be integrated.

Temporal ambiguity resolution The lexical/phonetic identity of a segment can be determined by acoustic features that arrive after the segment in question. pb rown The ambiguous first consonant of is clearly a /b/ after hearing ”rown” Thus, like in higher level language comprehension, temporal ambiguity resolution is an important issue.

Temporal ambiguity resolution Lexical/Phonetic Temporal Ambiguity can be caused by Vowel length (cue to speaking rate and stress) Lexical/Statistical effects Embedded words Subphonemic sensitivity can minimize or eliminate the effects of temporary phonetic ambiguity by Storing how ambiguous a segment is Keeping competitors active until resolution occurs.

Experiment 1: Effect of Time? How long does the gradient sensitivity to VOT remain? Need to examine: the effect of time on competitor fixations interaction with VOT

ms 5 ms 10 ms 15 ms VOT ms 25 ms 30 ms 35 ms 40 ms VOT Fixation proportion Time since word onset (ms) Experiment 1: Effect of time? Time course data suggests that gradiency is sticking around at least 1600 milliseconds after syllable onset. Response=

Experiment 1: Effect of Time? Analysis: earlylate Trial 1 Trial 2 Trial 3 Trial 4 Trial 5 Trial 6 Trial 7 Trial 8 Randomly sorted trials into two groups (early and late). Early Late For each group, fixations from only 1 time-bin were used Early: ms Late: ms Ensures independence of data in each time-bin (since each trial only contributes to one)

Experiment 1: VOT x Time VOT (ms) Category Boundary Late ( ms) Early ( ms) Main effect of time/b/: p=.001***/p/: p=.0001**** Response= Looks to Fixation proportion Main effect of VOT/b/: p=.015*/p/: p=.001*** Linear Trend for VOT /b/: p=.022*/p/: p=.009** No Interactionp>.1 Looks to

Experiment 1: VOT x Time VOT (ms) Late ( ms) Early ( ms) Response= Looks to Fixation proportion Category Boundary Main effect of time/b/: p=.001***/p/: p=.0001**** Main effect of VOT/b/: p=.006**/p/: p=.013* Linear Trend for VOT /b/: p=.0012**/p/: p=.02** No Interactionp>.1

Finally some conclusions Lexical activation exhibits gradient effects of subphonemic (VOT) variation. Effect is robust and long-lasting—could potentially be very helpful for resolving temporal ambiguity and integrating information over time. Effect of subphonemic variation is stronger for competitors than targets.

Finally some conclusions Experimental task is crucial to see sensitivity: more responses + less metalinguistic = more gradiency. ID Functions influenced by type of stimuli (e.g. words/nonwords) as well as number of response alternatives. Realistic tasks = more gradient ID functions.

Finally some conclusions Subphonemic variation in VOT is not discarded It is not but signal.