Behavioral studies have shown that the ability to discriminate between non-native speech sounds improves after seeing how the sounds are articulated. This study examined the influence of visual articulatory information on the neural correlates of non-native speech sound discrimination. English speakers' discrimination of the Hindi dental and retroflex sounds was measured using the mismatch negativity (MMN) event-related potential, before and after they completed one of three 8-min training conditions. In an audio-visual speech training condition ( = 14), each sound was presented with its corresponding visual articulation. In one control condition ( = 14), both sounds were presented with the same visual articulation, resulting in one congruent and one incongruent audio-visual pairing. In another control condition ( = 14), both sounds were presented with the same image of a still face. The control conditions aimed to rule out the possibility that the MMN is influenced by non-specific audio-visual pairings, or by general exposure to the dental and retroflex sounds over the course of the study. The results showed that audio-visual speech training reduced the latency of the MMN but did not affect MMN amplitude. No change in MMN amplitude or latency was observed for the two control conditions. The pattern of results suggests that a relatively short audio-visual speech training session (i.e., 8 min) may increase the speed with which the brain processes non-native speech sound contrasts. The absence of a training effect on MMN amplitude suggests a single session of audio-visual speech training does not lead to the formation of more discrete memory traces for non-native speech sounds. Longer and/or multiple sessions might be needed to influence the MMN amplitude.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7019039PMC
http://dx.doi.org/10.3389/fnhum.2020.00025DOI Listing

Publication Analysis

Top Keywords

non-native speech
20
audio-visual speech
16
speech training
16
mmn amplitude
16
speech sound
12
speech
9
visual articulatory
8
articulatory neural
8
neural correlates
8
correlates non-native
8

Similar Publications

Background: Language difficulties are common in autism spectrum disorder (ASD), a neurodevelopmental condition characterized by impairments in social communication as well as restricted and repetitive behaviors. Amongst infant siblings of children with an ASD diagnosis - who are at higher likelihood for developing ASD - a high proportion also show difficulties and delays in language acquisition.

Methods: In this study, we used functional magnetic resonance imaging (fMRI) to examine atypicalities associated with language processing in 9-month-old infants at high (HL) and typical (TL) familial likelihood for ASD.

View Article and Find Full Text PDF
Article Synopsis
  • Young infants initially can differentiate sounds from various languages, supporting the perceptual narrowing hypothesis, which suggests they become less sensitive to non-native phonemes as they grow.
  • This study investigated how 4-6 month-old Korean and Japanese infants respond to specific Thai phoneme contrasts, specifically looking at their ability to discriminate between different stop sounds based on voice onset time (VOT).
  • Findings revealed that Korean infants were sensitive to the pre-voiced vs. voiceless contrast, while Japanese infants were better at distinguishing the voiceless vs. voiceless aspirated sounds, highlighting significant differences in language input's impact on infants' phoneme discrimination abilities.
View Article and Find Full Text PDF

Effects of speaking style and semantic predictability on vowel production.

J Acoust Soc Am

November 2024

Department of Linguistics, Ohio State University, Columbus, Ohio 43210, USA.

Listener-oriented accounts of phonetic enhancement propose that talkers produce enhanced vowels to increase clarity when their interlocutor might experience communicative difficulty, e.g., for non-native interlocutors or for an unpredictable word given the semantic context.

View Article and Find Full Text PDF

Infants prefer infant-directed (ID) speech. Concerning talking faces, previous research showed that 3- and 5-month-olds prefer faces that produce native ID than native adult-directed (AD) speech, regardless of background speech being ID, AD or silent. Here, we explored whether infants also show a preference for non-native ID speech.

View Article and Find Full Text PDF

The developmental trajectories of tone perception among tone and non-tone language learning infants have received wide attention and discussion in recent decades under the perceptual attunement framework. Nevertheless, tone perception in infants from pitch accent and bilingual language backgrounds has not been well understood. The present study examined monolingual and bilingual Norwegian-learning infants' discrimination of two Cantonese tone contrasts at 5 and 10 months, ages corresponding to the onset and offset of perceptual attunement.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!