Brain-machine interfaces (BMIs) have great potential for applications that restore and assist communication for paralyzed individuals. Recently, BMIs decoding speech have gained considerable attention due to their potential for high information transfer rates. In this study, we propose a novel decoding approach based on hidden Markov models (HMMs) that uses the timing of neural signal changes to decode speech. We tested the decoder's performance by predicting vowels from electrocorticographic (ECoG) data of three human subjects. Our results show that timing-based features of ECoG signals are informative of vowel production and enable decoding accuracies significantly above the level of chance. This suggests that leveraging the temporal structure of neural activity to decode speech could play an important role towards developing highperformance, robust speech BMIs.

Download full-text PDF

Source
http://dx.doi.org/10.1109/EMBC.2016.7591002DOI Listing

Publication Analysis

Top Keywords

decoding speech
8
timing neural
8
neural signal
8
decode speech
8
decoding
4
speech timing
4
signal modulation
4
modulation brain-machine
4
brain-machine interfaces
4
interfaces bmis
4

Similar Publications

Decoding speech intent from non-frontal cortical areas.

J Neural Eng

January 2025

Department of Neurology, Northwestern University Feinberg School of Medicine, 320 East Superior St, Chicago, IL 60611, USA, Chicago, Illinois, 60611, UNITED STATES.

Brain-machine interfaces (BMIs) have advanced greatly in decoding speech signals originating from the speech motor cortices. Primarily, these BMIs target individuals with intact speech motor cortices but who are paralyzed by disrupted connections between frontal cortices and their articulators due to brainstem stroke or motor neuron diseases such as amyotrophic lateral sclerosis. A few studies have shown some information outside the speech motor cortices, such as in parietal and temporal lobes, that also may be useful for BMIs.

View Article and Find Full Text PDF

Real-world implementation of brain-computer interfaces (BCI) for continuous control of devices should ideally rely on fully asynchronous decoding approaches. That is, the decoding algorithm should continuously update its output by estimating the user's intended actions from real-time neural activity, without the need for any temporal alignment to an external cue. This kind of open-ended temporal flexibility is necessary to achieve naturalistic and intuitive control, but presents a challenge: how do we know when it is appropriate to decode anything at all? Activity in motor cortex is dynamic and modulates with many different types of actions (proximal arm control, hand control, speech, etc.

View Article and Find Full Text PDF

Affective voice signaling has significant biological and social relevance across various species, and different affective signaling types have emerged through the evolution of voice communication. These types range from basic affective voice bursts and nonverbal affective up to affective intonations superimposed on speech utterances in humans in the form of paraverbal prosodic patterns. These different types of affective signaling should have evolved to be acoustically and perceptually distinctive, allowing accurate and nuanced affective communication.

View Article and Find Full Text PDF

The extraction and analysis of pitch underpin speech and music recognition, sound segregation, and other auditory tasks. Perceptually, pitch can be represented as a helix composed of two factors: height monotonically aligns with frequency, while chroma cyclically repeats at doubled frequencies. Although the early perceptual and neurophysiological mechanisms for extracting pitch from acoustic signals have been extensively investigated, the equally essential subsequent stages that bridge to high-level auditory cognition remain less well understood.

View Article and Find Full Text PDF

Visual semantic decoding aims to extract perceived semantic information from the visual responses of the human brain and convert it into interpretable semantic labels. Although significant progress has been made in semantic decoding across individual visual cortices, studies on the semantic decoding of the ventral and dorsal cortical visual pathways remain limited. This study proposed a graph neural network (GNN)-based semantic decoding model on a natural scene dataset (NSD) to investigate the decoding differences between the dorsal and ventral pathways in process various parts of speech, including verbs, nouns, and adjectives.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!