Developmental changes in suprasegmental tonal duration were investigated in monolingual Mandarin-speaking children. Tone durations were acoustically measured in five- and eight-year-old children and adults. Children's tone duration and variability decreased with age.
View Article and Find Full Text PDFIt is widely known that language influences the way speech sounds are categorized. However, categorization of speech sounds by bilinguals is not well understood. There is evidence that bilinguals have different category boundaries than monolinguals, and there is evidence suggesting that bilinguals' phonemic boundaries can shift with language context.
View Article and Find Full Text PDFSeveral fixed classification experiments test the hypothesis that F(1), f(0), and closure voicing covary between intervocalic stops contrasting for [voice] because they integrate perceptually. The perceptual property produced by the integration of these acoustic properties was at first predicted to be the presence of low frequency energy in the vicinity of the stop, which is considerable in [+voice] stops but slight in [-voice] stops. Both F(1) and f(0) at the edges of vowels flanking the stop were found to integrate perceptually with the continuation of voicing into the stop, but not to integrate with one another.
View Article and Find Full Text PDFPhilos Trans R Soc Lond B Biol Sci
March 2008
Speech perception is remarkably robust. This paper examines how acoustic and auditory properties of vowels and consonants help to ensure intelligibility. First, the source-filter theory of speech production is briefly described, and the relationship between vocal-tract properties and formant patterns is demonstrated for some commonly occurring vowels.
View Article and Find Full Text PDFThis study was designed to test the iambic/trochaic law, which claims that elements contrasting in duration naturally form rhythmic groupings with final prominence, whereas elements contrasting in intensity form groupings with initial prominence. It was also designed to evaluate whether the iambic/trochaic law describes general auditory biases, or whether rhythmic grouping is speech or language specific. In two experiments, listeners were presented with sequences of alternating /ga/ syllables or square wave segments that varied in either duration or intensity and were asked to indicate whether they heard a trochaic (i.
View Article and Find Full Text PDFThe present study examined several potential distinctiveness-enhancing correlates of vowels produced in utterance focus by talkers of American English, French, and German. These correlates included possible increases in vowel space size, in formant movement within individual vowels, and in duration variance among vowels. Each language group enhanced the distinctiveness of vowels in [+focus] context but used somewhat differing means to achieve this.
View Article and Find Full Text PDFAuditory pitch patterns are significant ecological features to which nervous systems have exquisitely adapted. Pitch patterns are found embedded in many contexts, enabling different information-processing goals. Do the psychological functions of pitch patterns determine the neural mechanisms supporting their perception, or do all pitch patterns, regardless of function, engage the same mechanisms? This issue is pursued in the present study by using 150-water positron emission tomography to study brain activations when two subject groups discriminate pitch patterns in their respective native languages, one of which is a tonal language and the other of which is not.
View Article and Find Full Text PDFJ Acoust Soc Am
September 2004
Behavioral experiments with infants, adults, and nonhuman animals converge with neurophysiological findings to suggest that there is a discontinuity in auditory processing of stimulus components differing in onset time by about 20 ms. This discontinuity has been implicated as a basis for boundaries between speech categories distinguished by voice onset time (VOT). Here, it is investigated how this discontinuity interacts with the learning of novel perceptual categories.
View Article and Find Full Text PDFThis chapter focuses on one of the first steps in comprehending spoken language: How do listeners extract the most fundamental linguistic elements-consonants and vowels, or the distinctive features which compose them-from the acoustic signal? We begin by describing three major theoretical perspectives on the perception of speech. Then we review several lines of research that are relevant to distinguishing these perspectives. The research topics surveyed include categorical perception, phonetic context effects, learning of speech and related nonspeech categories, and the relation between speech perception and production.
View Article and Find Full Text PDFJ Speech Lang Hear Res
April 2003
Inter- and intratalker variation in the production of lexical tones may contribute to acoustic overlap among tone categories. The present study investigated whether such category overlap gives rise to perceptual ambiguity and, if so, whether listeners are able to reduce this ambiguity using contextual information. In the first experiment, native Cantonese-speaking listeners were asked to identify isolated Cantonese level tones produced by 7 talkers.
View Article and Find Full Text PDFEvoked magnetic fields were recorded from 18 adult volunteers using magnetoencephalography (MEG) during perception of speech stimuli (the endpoints of a voice onset time (VOT) series ranging from /ga/ to /ka/), analogous nonspeech stimuli (the endpoints of a two-tone series varying in relative tone onset time (TOT), and a set of harmonically complex tones varying in pitch. During the early time window (approximately 60 to approximately 130 ms post-stimulus onset), activation of the primary auditory cortex was bilaterally equal in strength for all three tasks. During the middle (approximately 130 to 800 ms) and late (800 to 1400 ms) time windows of the VOT task, activation of the posterior portion of the superior temporal gyrus (STGp) was greater in the left hemisphere than in the right hemisphere, in both group and individual data.
View Article and Find Full Text PDFTwelve male listeners categorized 54 synthetic vowel stimuli that varied in second and third formant frequency on a Bark scale into the American English vowel categories [see text]. A neuropsychologically plausible model of categorization in the visual domain, the Striatal Pattern Classifier (SPC; Ashby & Waldron, 1999), is generalized to the auditory domain and applied separately to the data from each observer. Performance of the SPC is compared with that of the successful Normal A Posteriori Probability model (NAPP; Nearey, 1990; Nearey & Hogan, 1986) of auditory categorization.
View Article and Find Full Text PDFPhilos Trans R Soc Lond B Biol Sci
April 2002
In recent years, there has been much interest in characterizing statistical properties of natural stimuli in order to better understand the design of perceptual systems. A fruitful approach has been to compare the processing of natural stimuli in real perceptual systems with that of ideal observers derived within the framework of Bayesian statistical decision theory. While this form of optimization theory has provided a deeper understanding of the information contained in natural stimuli as well as of the computational principles employed in perceptual systems, it does not directly consider the process of natural selection, which is ultimately responsible for design.
View Article and Find Full Text PDFVowel nuclei of syllables appear to provide a relatively stable (although not stationary) frame of reference for judging consonant events. We offer evidence that reliable consonant identification demands prior or simultaneous evaluation of this "vocalic frame." Listeners were presented a list of /bVs/, /dVs/, and /gVs/ syllables and were instructed to press a response key immediately upon recognizing a particular initial consonant target.
View Article and Find Full Text PDF