Healthcare sensors represent a valid and non-invasive instrument to capture and analyse physiological data. Several vital signals, such as voice signals, can be acquired anytime and anywhere, achieved with the least possible discomfort to the patient thanks to the development of increasingly advanced devices. The integration of sensors with artificial intelligence techniques contributes to the realization of faster and easier solutions aimed at improving early diagnosis, personalized treatment, remote patient monitoring and better decision making, all tasks vital in a critical situation such as that of the COVID-19 pandemic. This paper presents a study about the possibility to support the early and non-invasive detection of COVID-19 through the analysis of voice signals by means of the main machine learning algorithms. If demonstrated, this detection capacity could be embedded in a powerful mobile screening application. To perform this important study, the Coswara dataset is considered. The aim of this investigation is not only to evaluate which machine learning technique best distinguishes a healthy voice from a pathological one, but also to identify which vowel sound is most seriously affected by COVID-19 and is, therefore, most reliable in detecting the pathology. The results show that Random Forest is the technique that classifies most accurately healthy and pathological voices. Moreover, the evaluation of the vowel /e/ allows the detection of the effects of COVID-19 on voice quality with a better accuracy than the other vowels.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC8500467 | PMC |
http://dx.doi.org/10.1007/s13369-021-06041-4 | DOI Listing |
Ann Fam Med
January 2025
Departments of Psychiatry and Emergency Medicine, University of Arkansas for Medical Sciences, Little Rock, Arkansas.
Purpose: Mental health screening is recommended by the US Preventive Services Task Force for all patients in areas where treatment options are available. Still, it is estimated that only 4% of primary care patients are screened for depression. The goal of this study was to evaluate the efficacy of machine learning technology (Kintsugi Voice, v1, Kintsugi Mindful Wellness, Inc) to detect and analyze voice biomarkers consistent with moderate to severe depression, potentially allowing for greater compliance with this critical primary care public health need.
View Article and Find Full Text PDFEmotion
January 2025
Department of Psychology, Cognitive and Affective Neuroscience Unit, University of Zurich.
Affective voice signaling has significant biological and social relevance across various species, and different affective signaling types have emerged through the evolution of voice communication. These types range from basic affective voice bursts and nonverbal affective up to affective intonations superimposed on speech utterances in humans in the form of paraverbal prosodic patterns. These different types of affective signaling should have evolved to be acoustically and perceptually distinctive, allowing accurate and nuanced affective communication.
View Article and Find Full Text PDFJ Spinal Cord Med
January 2025
Speech-Language-Hearing Sciences, Medical School, Federal University of Minas Gerais, Belo Horizonte, Brazil.
Introduction: Spinal cord injury is a physiological disruption often caused by trauma, leading to severe physical and psychological effects, including irreversible impairment and disability. Cervical injuries, particularly between C1 and C8, are the most severe, potentially causing diaphragm paralysis and requiring mechanical ventilation. Reduced respiratory muscle strength not only affects respiratory function but also significantly impacts voice, speech, and communication, which are crucial for quality of life.
View Article and Find Full Text PDFCortex
December 2024
Institute of Research in Psychology (IPSY) & Institute of Neuroscience (IoNS), Louvain Bionics Center, University of Louvain (UCLouvain), Louvain-la-Neuve, Belgium; School of Health Sciences, HES-SO Valais-Wallis, The Sense Innovation and Research Center, Lausanne & Sion, Switzerland. Electronic address:
Effective social communication depends on the integration of emotional expressions coming from the face and the voice. Although there are consistent reports on how seeing and hearing emotion expressions can be automatically integrated, direct signatures of multisensory integration in the human brain remain elusive. Here we implemented a multi-input electroencephalographic (EEG) frequency tagging paradigm to investigate neural populations integrating facial and vocal fearful expressions.
View Article and Find Full Text PDFJ Child Lang
January 2025
Department of Developmental Neuroscience, IRCCS Stella Maris Foundation, Pisa, Italy.
To investigate how a high risk for infant neurological impairment affects the quality of infant verbal interactions, and in particular properties of infant-directed speech, spontaneous interactions between 14 mothers and their 4.5-month-old infants at high risk for neurological disorders (7 female) were recorded and acoustically compared with those of 14 dyads with typically developing infants (8 female). Mothers of at-risk infants had proportionally less voicing, and the proportion of voicing decreased with increasing severity of the infants' long-term outcome.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!