We investigated how neural oscillations code the hierarchical nature of stress rhythms in speech and how stress processing varies with language experience. By measuring phase synchrony of multilevel EEG-acoustic tracking and intra-brain cross-frequency coupling, we show the encoding of stress involves different neural signatures (delta rhythms = stress foot rate; theta rhythms = syllable rate), is stronger for amplitude vs. duration stress cues, and induces nested delta-theta coherence mirroring the stress-syllable hierarchy in speech.
View Article and Find Full Text PDFSurrounding context influences speech listening, resulting in dynamic shifts to category percepts. To examine its neural basis, event-related potentials (ERPs) were recorded during vowel identification with continua presented in random, forward, and backward orders to induce perceptual warping. Behaviorally, sequential order shifted individual listeners' categorical boundary, versus random delivery, revealing perceptual warping (biasing) of the heard phonetic category dependent on recent stimulus history.
View Article and Find Full Text PDFProc ACM Interact Mob Wearable Ubiquitous Technol
December 2020
Stressful conversation is a frequently occurring stressor in our daily life. Stressors not only adversely affect our physical and mental health but also our relationships with family, friends, and coworkers. In this paper, we present a model to automatically detect stressful conversations using wearable physiological and inertial sensors.
View Article and Find Full Text PDFPurpose This study measures the experience of spontaneous speech in everyday speaking situations. Spontaneity of speech is a novel concept developed to account for the subjective experience of speaking. Spontaneous speech is characterized by little premeditation and effortless production, and it is enjoyable and meaningful.
View Article and Find Full Text PDFHow did vocal language originate? Before trying to determine how referential vocabulary or syntax may have arisen, it is critical to explain how ancient hominins began to produce vocalization flexibly, without binding to emotions or functions. A crucial factor in the vocal communicative split of hominins from the ape background may thus have been copious, functionally flexible vocalization, starting in infancy and continuing throughout life, long before there were more advanced linguistic features such as referential vocabulary. 2-3 month-old modern human infants produce "protophones", including at least three types of functionally flexible non-cry precursors to speech rarely reported in other ape infants.
View Article and Find Full Text PDFPrior research has not evaluated acoustic features contributing to perception of human infant vocal distress or lack thereof on a continuum. The present research evaluates perception of infant vocalizations along a continuum ranging from the most prototypical intensely distressful cry sounds ("wails") to the most prototypical of infant sounds that typically express no distress (non-distress "vocants"). Wails are deemed little if at all related to speech while vocants are taken to be clear precursors to speech.
View Article and Find Full Text PDFPurpose: This article examines cepstral/spectral analyses of sustained /α/ vowels produced by speakers with hypokinetic dysarthria secondary to idiopathic Parkinson's disease (PD) before and after Lee Silverman Voice Treatment (LSVT®LOUD) and the relationship of these measures with overall voice intensity.
Methodology: Nine speakers with PD were examined in a pre-/post-treatment design, with multiple daily audio recordings before and after treatment. Sustained vowels were analyzed for cepstral peak prominence (CPP), CPP standard deviation (CPP SD), low/high spectral ratio (L/H SR), and Cepstral/Spectral Index of Dysphonia (CSID) using the KAYPENTAX computer software.
Proc ACM Interact Mob Wearable Ubiquitous Technol
March 2018
Monitoring of in-person conversations has largely been done using acoustic sensors. In this paper, we propose a new method to detect moment-by-moment conversation episodes by analyzing breathing patterns captured by a mobile respiration sensor. Since breathing is affected by physical and cognitive activities, we develop a comprehensive method for cleaning, screening, and analyzing noisy respiration data captured in the field environment at individual breath cycle level.
View Article and Find Full Text PDFThe primary vocal registers of modal, falsetto, and fry have been studied in adults but not per se in infancy. The vocal ligament is thought to play a critical role in the modal-falsetto contrast but is still developing during infancy (Tateya and Tateya, 2015). Cover tissues are also implicated in the modal-fry contrast, but the low fundamental frequency (f) cutoff of 70 Hz, shared between genders, suggests a psychoacoustic basis for the contrast.
View Article and Find Full Text PDFDebates about neonatal imitation remain more open than Keven & Akins (K&A) imply. K&A do not recognize the primacy of the question concerning differential imitation and the links between experimental designs and more or less plausible theoretical assumptions. Moreover, they do not acknowledge previous theorizing on spontaneous behavior, the explanatory power of entrainment, and subtle connections with social cognition.
View Article and Find Full Text PDFNeonatal imitation has rich implications for neuroscience, developmental psychology, and social cognition, but there is little consensus about this phenomenon. The primary empirical question, whether or not neonatal imitation exists, is not settled. Is it possible to give a balanced evaluation of the theories and methodologies at stake so as to facilitate real progress with respect to the primary empirical question? In this paper, we address this question.
View Article and Find Full Text PDFProc ACM Int Conf Ubiquitous Comput
September 2016
Differential privacy concepts have been successfully used to protect anonymity of individuals in population-scale analysis. Sharing of mobile sensor data, especially physiological data, raise different privacy challenges, that of protecting private behaviors that can be revealed from time series of sensor data. Existing privacy mechanisms rely on noise addition and data perturbation.
View Article and Find Full Text PDFProc Natl Acad Sci U S A
April 2013
We report on the emergence of functional flexibility in vocalizations of human infants. This vastly underappreciated capability becomes apparent when prelinguistic vocalizations express a full range of emotional content--positive, neutral, and negative. The data show that at least three types of infant vocalizations (squeals, vowel-like sounds, and growls) occur with this full range of expression by 3-4 mo of age.
View Article and Find Full Text PDFVocal motor development in infancy provides a crucial foundation for language development. Some significant early accomplishments include learning to control the process of phonation (the production of sound at the larynx) and learning to produce the sounds of one's language. Previous work has shown that social reinforcement shapes the kinds of vocalizations infants produce.
View Article and Find Full Text PDFPurpose: The prelinguistic infant's babbling repertoire of syllables--the phonological categories that form the basis for early word learning--is noticed by caregivers who interact with infants around them. Prior research on babbling has not explored the caregiver's role in recognition of early vocal categories as foundations for word learning. In the present work, the authors begin to address this gap.
View Article and Find Full Text PDFJ Speech Lang Hear Res
August 2011
Purpose: This study explored whether breathing behaviors of infants within the 2nd year of life differ between tidal breathing and breathing supporting single unarticulated syllables and canonical/articulated syllables.
Method: Vocalizations and breathing kinematics of 9 infants between 53 and 90 weeks of age were recorded. A strict selection protocol was used to identify analyzable breath cycles.
Acoustic analysis of infant vocalizations has typically employed traditional acoustic measures drawn from adult speech acoustics, such as f(0), duration, formant frequencies, amplitude, and pitch perturbation. Here an alternative and complementary method is proposed in which data-derived spectrographic features are central. 1-s-long spectrograms of vocalizations produced by six infants recorded longitudinally between ages 3 and 11 months are analyzed using a neural network consisting of a self-organizing map and a single-layer perceptron.
View Article and Find Full Text PDFHesitations have been considered to serve both cognitive and linguistic functions. This study presents analyses of children's hesitations while producing English derived words with the suffix -ity. Two questions were considered: Do children's linguistic skills influence their use and frequency of hesitations when producing derived words, and do children's use of hesitations vary as a function of word frequency.
View Article and Find Full Text PDFInfant phonation is highly variable in many respects, including the basic vibratory patterns by which the vocal tissues create acoustic signals. Previous studies have identified the regular occurrence of nonmodal phonation types in normal infant phonation. The glottis is like many oscillating systems that, because of nonlinear relationships among the elements, may vibrate in ways representing the deterministic patterns classified theoretically within the mathematical framework of nonlinear dynamics.
View Article and Find Full Text PDFSpectral amplitude measures are sensitive to varying degrees of vocal fold adduction in normal speakers. This study examined the applicability of harmonic amplitude differences to adductor spasmodic dysphonia (ADSD) in comparison with normal controls. Amplitudes of the first and second harmonics (H1, H2) and of harmonics affiliated with the first, second, and third formants (A1, A2, A3) were obtained from spectra of vowels and /i/ excerpted from connected speech.
View Article and Find Full Text PDFVoice clinicians require an objective, reliable, and relatively automatic method to assess voice change after medical, surgical, or behavioral intervention. This measure must be sensitive to a variety of voice qualities and severities, and preferably should reflect voice in continuous speech. The long-term average spectrum (LTAS) is a fast Fourier transform-generated power spectrum whose properties can be compared with a Gaussian bell curve using spectral moments analysis.
View Article and Find Full Text PDF