Many computational theories have been developed to improve artificial phonetic classification performance from linguistic auditory streams. However, less attention has been given to psycholinguistic data and neurophysiological features recently found in cortical tissue. We focus on a context in which basic linguistic units-such as phonemes-are extracted and robustly classified by humans and other animals from complex acoustic streams in speech data. We are especially motivated by the fact that 8-month-old human infants can accomplish segmentation of words from fluent audio streams based exclusively on the statistical relationships between neighboring speech sounds without any kind of supervision. In this paper, we introduce a biologically inspired and fully unsupervised neurocomputational approach that incorporates key neurophysiological and anatomical cortical properties, including columnar organization, spontaneous micro-columnar formation, adaptation to contextual activations and Sparse Distributed Representations (SDRs) produced by means of partial N-Methyl-D-aspartic acid (NMDA) depolarization. Its feature abstraction capabilities show promising phonetic invariance and generalization attributes. Our model improves the performance of a Support Vector Machine (SVM) classifier for monosyllabic, disyllabic and trisyllabic word classification tasks in the presence of environmental disturbances such as white noise, reverberation, and pitch and voice variations. Furthermore, our approach emphasizes potential self-organizing cortical principles achieving improvement without any kind of optimization guidance which could minimize hypothetical loss functions by means of-for example-backpropagation. Thus, our computational model outperforms multiresolution spectro-temporal auditory feature representations using only the statistical sequential structure immerse in the phonotactic rules of the input stream.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC6555517 | PMC |
http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0217966 | PLOS |
Ear Hear
December 2024
Department of Audiology and Phoniatrics, Charité - Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Berlin, Germany.
Language outcomes of children with hearing loss remain heterogeneous despite recent advances in treatment and intervention. Consonants with high frequency, in particular, continue to pose challenges to affected children's speech perception and production. In this review, the authors evaluate findings of how enriched child-directed speech and song might function as a form of early family-centered intervention to remedy the effects of hearing loss on consonant acquisition already during infancy.
View Article and Find Full Text PDFNeuroscience
December 2024
School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA.
This study assessed the neural mechanisms and relative saliency of categorization for speech sounds and comparable graphemes (i.e., visual letters) of the same phonetic label.
View Article and Find Full Text PDFCognition
January 2025
École d'orthophonie et d'audiologie, l'Université de Montréal, Canada.
Interactions between the context in which a sensorimotor skill is learned and the recall of that memory have been primarily studied in limb movements, but speech production requires movement, and many aspects of speech processing are influenced by task-relevant contextual information. Here, in ecologically valid speech (read sentences), we test whether English-French bilinguals can use the language of production to acquire and recall distinct motor plans for similar speech sounds spanning the production workspace. Participants experienced real-time alterations of auditory feedback while producing interleaved English and French sentences.
View Article and Find Full Text PDFInfancy
December 2024
Department of Linguistics, University of Potsdam, Potsdam, Brandenburg, Germany.
J Exp Child Psychol
January 2025
Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, UK.
Language processing encompasses a sophisticated interplay of phonological (sound-based) and semantic (meaning-based) processes. This intricate interaction develops progressively during early language acquisition. It involves not only the addition of new words to the child's vocabulary but also the evolving organization of lexico-semantic networks.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!