Brain-machine interfaces (BMIs) have great potential for applications that restore and assist communication for paralyzed individuals. Recently, BMIs decoding speech have gained considerable attention due to their potential for high information transfer rates. In this study, we propose a novel decoding approach based on hidden Markov models (HMMs) that uses the timing of neural signal changes to decode speech. We tested the decoder's performance by predicting vowels from electrocorticographic (ECoG) data of three human subjects. Our results show that timing-based features of ECoG signals are informative of vowel production and enable decoding accuracies significantly above the level of chance. This suggests that leveraging the temporal structure of neural activity to decode speech could play an important role towards developing highperformance, robust speech BMIs.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/EMBC.2016.7591002 | DOI Listing |
J Neural Eng
January 2025
Department of Neurology, Northwestern University Feinberg School of Medicine, 320 East Superior St, Chicago, IL 60611, USA, Chicago, Illinois, 60611, UNITED STATES.
Brain-machine interfaces (BMIs) have advanced greatly in decoding speech signals originating from the speech motor cortices. Primarily, these BMIs target individuals with intact speech motor cortices but who are paralyzed by disrupted connections between frontal cortices and their articulators due to brainstem stroke or motor neuron diseases such as amyotrophic lateral sclerosis. A few studies have shown some information outside the speech motor cortices, such as in parietal and temporal lobes, that also may be useful for BMIs.
View Article and Find Full Text PDFJ Neural Eng
January 2025
University of Pittsburgh, 1622 Locust St, Pittsburgh, Pennsylvania, 15219, UNITED STATES.
Real-world implementation of brain-computer interfaces (BCI) for continuous control of devices should ideally rely on fully asynchronous decoding approaches. That is, the decoding algorithm should continuously update its output by estimating the user's intended actions from real-time neural activity, without the need for any temporal alignment to an external cue. This kind of open-ended temporal flexibility is necessary to achieve naturalistic and intuitive control, but presents a challenge: how do we know when it is appropriate to decode anything at all? Activity in motor cortex is dynamic and modulates with many different types of actions (proximal arm control, hand control, speech, etc.
View Article and Find Full Text PDFEmotion
January 2025
Department of Psychology, Cognitive and Affective Neuroscience Unit, University of Zurich.
Affective voice signaling has significant biological and social relevance across various species, and different affective signaling types have emerged through the evolution of voice communication. These types range from basic affective voice bursts and nonverbal affective up to affective intonations superimposed on speech utterances in humans in the form of paraverbal prosodic patterns. These different types of affective signaling should have evolved to be acoustically and perceptually distinctive, allowing accurate and nuanced affective communication.
View Article and Find Full Text PDFJ Neurosci
January 2025
Department of Psychology, Chinese University of Hong Kong, Hong Kong SAR, China
The extraction and analysis of pitch underpin speech and music recognition, sound segregation, and other auditory tasks. Perceptually, pitch can be represented as a helix composed of two factors: height monotonically aligns with frequency, while chroma cyclically repeats at doubled frequencies. Although the early perceptual and neurophysiological mechanisms for extracting pitch from acoustic signals have been extensively investigated, the equally essential subsequent stages that bridge to high-level auditory cognition remain less well understood.
View Article and Find Full Text PDFInt J Neural Syst
January 2025
Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou 313001, P. R. China.
Visual semantic decoding aims to extract perceived semantic information from the visual responses of the human brain and convert it into interpretable semantic labels. Although significant progress has been made in semantic decoding across individual visual cortices, studies on the semantic decoding of the ventral and dorsal cortical visual pathways remain limited. This study proposed a graph neural network (GNN)-based semantic decoding model on a natural scene dataset (NSD) to investigate the decoding differences between the dorsal and ventral pathways in process various parts of speech, including verbs, nouns, and adjectives.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!