Background: Automated analysis of natural speech is emerging as a promising digital biomarker of Alzheimer's disease (AD). As speech is a complex process, relying on multiple interacting cognitive functions, fine-grained analysis of speech may have the potential to capture subtle cognitive deficits in the very early stages of AD. Here, we examined the association between amyloid-beta (Aβ) pathology and acoustic speech characteristics in a group of cognitively normal Dutch adults.

Method: We included 50 cognitively normal (n = 23 Aβ-positive, 46%) older adults from three clinical cohorts at the Alzheimer Center Amsterdam (Table 1). Aβ-status was based on local cut-offs for Aβ-concentrations in cerebrospinal-fluid or visual inspection of amyloid positron emission tomography (PET)-imaging. In a five day remote burst assessment, seventeen tablet-based tasks (picture description, journal-prompt storytelling, verbal-fluency) were administered, with a daily administration time of approximately 10 minutes. Various acoustic features, including silent pauses, pause-to-word-ratio, and perceived loudness, were extracted from the voice recordings. Mean burst scores were calculated over the five-day period. The association between Aβ-pathology and acoustic features was investigated using linear regression models separately for each subtask.

Results: Pause-to-word ratio was higher in Aβ-positive individuals in the repetitive picture description subtask (β = -0.05, p = 0.04) and the journaling subtask (β = -0.07, p = 0.03) compared to the Aβ-negative individuals. Although statistical significance between groups was not reached for other acoustic features, a pattern was observed that the Aβ-positive group consistently displayed more medium pauses, longer audio duration, more local jitter, higher intensity variance, a higher fundamental frequency, and fewer long pauses in all multi-day subtasks (Figure 1).

Conclusion: Our results demonstrated that Aβ-pathology was associated with more pauses during free speech in cognitively healthy adults. This supports the notion that remote multi-day speech assessments have the potential to monitor cognitive changes, for example in observational studies and decentralized clinical trials in preclinical AD. Next steps include examining the association between Aβ-pathology and linguistic features using the same sample.

Download full-text PDF

Source
http://dx.doi.org/10.1002/alz.087304DOI Listing

Publication Analysis

Top Keywords

acoustic features
12
cognitively normal
8
picture description
8
association aβ-pathology
8
speech
6
clinical manifestations
4
manifestations background
4
background automated
4
automated analysis
4
analysis natural
4

Similar Publications

Salient Voice Symptoms in Primary Muscle Tension Dysphonia.

J Voice

January 2025

School of Behavioral and Brain Sciences, Department of Speech, Language, and Hearing, Callier Center for Communication Disorders, University of Texas at Dallas, Richardson, TX; Department of Otolaryngology - Head and Neck Surgery, University of Texas Southwestern Medical Center, Dallas, TX. Electronic address:

Introduction: Patients with primary muscle tension dysphonia (pMTD) commonly report symptoms of vocal effort, fatigue, discomfort, odynophonia, and aberrant vocal quality (eg, vocal strain, hoarseness). However, voice symptoms most salient to pMTD have not been identified. Furthermore, how standard vocal fatigue and vocal tract discomfort indices that capture persistent symptoms-like the Vocal Fatigue Index (VFI) and Vocal Tract Discomfort Scale (VTDS)-relate to acute symptoms experienced at the time of the voice evaluation is unclear.

View Article and Find Full Text PDF

Machine Learning-Based Estimation of Hoarseness Severity Using Acoustic Signals Recorded During High-Speed Videoendoscopy.

J Voice

January 2025

Division of Phoniatrics and Pediatric Audiology at the Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital Erlangen, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91054 Erlangen, Germany.

Objectives: This study investigates the use of sustained phonations recorded during high-speed videoendoscopy (HSV) for machine learning-based assessment of hoarseness severity (H). The performance of this approach is compared with conventional recordings obtained during voice therapy to evaluate key differences and limitations of HSV-derived acoustic recordings.

Methods: A database of 617 voice recordings with a duration of 250 ms was gathered during HSV examination (HS).

View Article and Find Full Text PDF

Genetic and audiological determinants of hearing loss in high-risk neonates.

Braz J Otorhinolaryngol

January 2025

Shanghai Jiao Tong University, School of Medicine, Hainan Branch of Shanghai Children's Medical Center, Department of Otorhinolaryngology, Sanya, China; Shanghai Jiao Tong University, School of Medicine, Shanghai Children's Medical Center, Department of Otorhinolaryngology, Shanghai, China. Electronic address:

Objective: We aimed to investigate the correlation between prevalent risk factors for high-risk neonates in neonatal intensive care unit and their hearing loss, and to examine the audiological features and genetic profiles associated with different deafness mutations in our tertiary referral center. This research seeks to deepen our understanding of the etiology behind congenital hearing loss.

Methods: We conducted initial hearing screenings, including automated auditory brainstem response, distortion product otoacoustic emission, and acoustic immittance on 443 high-risk neonates within 7 days after birth and 42 days (if necessary) after birth.

View Article and Find Full Text PDF

Prosodic Modifications to Challenging Communicative Environments in Preschoolers.

Lang Speech

January 2025

Department of Educational Psychology, Leadership, & Counseling, Texas Tech University, USA.

Adapting one's speaking style is particularly crucial as children start interacting with diverse conversational partners in various communication contexts. The study investigated the capacity of preschool children aged 3-5 years ( = 28) to modify their speaking styles in response to background noise, referred to as noise-adapted speech, and when talking to an interlocutor who pretended to have hearing loss, referred to as clear speech. We examined how two modified speaking styles differed across the age range.

View Article and Find Full Text PDF

A comprehensive analysis of everyday sound perception can be achieved using Electroencephalography (EEG) with the concurrent acquisition of information about the environment. While extensive research has been dedicated to speech perception, the complexities of auditory perception within everyday environments, specifically the types of information and the key features to extract, remain less explored. Our study aims to systematically investigate the relevance of different feature categories: discrete sound-identity markers, general cognitive state information, and acoustic representations, including discrete sound onset, the envelope, and mel-spectrogram.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!