Introduction: Speech BCIs aim at reconstructing speech in real time from ongoing cortical activity. Ideal BCIs would need to reconstruct speech audio signal frame by frame on a millisecond-timescale. Such approaches require fast computation. In this respect, linear decoder are good candidates and have been widely used in motor BCIs. Yet, they have been very seldomly studied for speech reconstruction, and never for reconstruction of articulatory movements from intracranial activity. Here, we compared vanilla linear regression, ridge-regularized linear regressions, and partial least squares regressions for offline decoding of overt speech from cortical activity.

Methods: Two decoding paradigms were investigated: (1) direct decoding of acoustic vocoder features of speech, and (2) indirect decoding of vocoder features through an intermediate articulatory representation chained with a real-time-compatible DNN-based articulatory-to-acoustic synthesizer. Participant's articulatory trajectories were estimated from an electromagnetic-articulography dataset using dynamic time warping. The accuracy of the decoders was evaluated by computing correlations between original and reconstructed features.

Results: We found that similar performance was achieved by all linear methods well above chance levels, albeit without reaching intelligibility. Direct and indirect methods achieved comparable performance, with an advantage for direct decoding.

Discussion: Future work will address the development of an improved neural speech decoder compatible with fast frame-by-frame speech reconstruction from ongoing activity at a millisecond timescale.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC10326283PMC
http://dx.doi.org/10.3389/fnhum.2023.1124065DOI Listing

Publication Analysis

Top Keywords

overt speech
8
cortical activity
8
linear methods
8
speech
8
speech reconstruction
8
vocoder features
8
decoding
5
linear
5
speech decoding
4
decoding cortical
4

Similar Publications

Private speech is a tool through which children self-regulate. The regulatory content of children's overt private speech is associated with response to task difficulty and task performance. Parenting is proposed to play a role in the development of private speech as co-regulatory interactions become represented by the child as private speech to regulate thinking and behaviour.

View Article and Find Full Text PDF
Article Synopsis
  • Drug-induced stuttering is an acquired speech disorder caused by certain medications, resembling developmental stuttering, and has been primarily studied through case reports and adverse drug reactions.
  • A recent study analyzed electronic health records from a major medical center to identify and classify drugs linked to this type of stuttering, reviewing 40 suspected cases.
  • The findings revealed that 18 different drugs were associated with stuttering in 22 individuals, especially in the classes of antiseizure agents, CNS stimulants, and antidepressants, with topiramate being the most commonly implicated drug; the study emphasizes the need for better documentation of medication-related speech issues in EHRs.
View Article and Find Full Text PDF

Sentence production is the uniquely human ability to transform complex thoughts into strings of words. Despite the importance of this process, language production research has primarily focused on single words. It remains an untested assumption that insights from this literature generalize to more naturalistic utterances like sentences.

View Article and Find Full Text PDF

Continuous and discrete decoding of overt speech with scalp electroencephalography (EEG).

J Neural Eng

October 2024

Electrical and Computer Engineering, University of Houston, N308 Engineering Building I, Houston, Texas, 77204-4005, UNITED STATES.

Article Synopsis
  • * This research investigates the use of non-invasive EEG to develop speech Brain-Computer Interfaces (BCIs) that decode speech features directly, aiming for a more natural communication method.
  • * Deep learning models, such as CNNs and RNNs, were tested for speech decoding tasks, showing significant success in distinguishing both discrete and continuous speech elements, while also indicating the importance of specific EEG frequency bands for performance.
View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!