Speechreading-gathering speech information from talkers' faces-supports speech perception when speech acoustics are degraded. Benefitting from speechreading, however, requires listeners to visually fixate talkers during face-to-face interactions. The purpose of this study is to test the hypothesis that preschool-aged children allocate their eye gaze to a talker when speech acoustics are degraded. We implemented a looking-while-listening paradigm to quantify children's eye gaze to an unfamiliar female talker and two images of familiar objects presented on a screen while the children listened to speech. We tested 31 children (12 girls), ages 26-48 months, who had normal hearing (NH group, n = 19) or bilateral sensorineural hearing loss and used hearing devices (D/HH group, n = 12). Children's eye gaze was video-recorded as the talker verbally labeled one of the images, either in quiet or in the presence of an unfamiliar two-talker male speech masker. Children's eye gaze to the target image, distractor image, and female talker was coded every 33 ms off-line by trained observers. Bootstrapped differences of time series (BDOTS) analyses and ternary plots were used to determine differences in visual fixations of the talker between listening conditions in the NH and D/HH groups. Results suggest that the NH group visually fixated the talker more in the masker condition than in quiet. We did not observe statistically discernable differences in visual fixations of the talker between the listening conditions for the D/HH group. Gaze patterns of the NH group in the masker condition looked like gaze patterns of the D/HH group.

Download full-text PDF

Source
http://dx.doi.org/10.3758/s13414-024-03001-6DOI Listing

Publication Analysis

Top Keywords

eye gaze
20
children's eye
16
d/hh group
12
talker
8
gaze talker
8
speech acoustics
8
acoustics degraded
8
female talker
8
differences visual
8
visual fixations
8

Similar Publications

Diagnosis of Parkinson's disease by eliciting trait-specific eye movements in multi-visual tasks.

J Transl Med

January 2025

School of Information and Communication Engineering, Dalian University of Technology, No. 2 Linggong Road, 116024, Dalian, China.

Background: Parkinson's Disease (PD) is a neurodegenerative disorder, and eye movement abnormalities are a significant symptom of its diagnosis. In this paper, we developed a multi-task driven by eye movement in a virtual reality (VR) environment to elicit PD-specific eye movement abnormalities. The abnormal features were subsequently modeled by using the proposed deep learning algorithm to achieve an auxiliary diagnosis of PD.

View Article and Find Full Text PDF

Speechreading-gathering speech information from talkers' faces-supports speech perception when speech acoustics are degraded. Benefitting from speechreading, however, requires listeners to visually fixate talkers during face-to-face interactions. The purpose of this study is to test the hypothesis that preschool-aged children allocate their eye gaze to a talker when speech acoustics are degraded.

View Article and Find Full Text PDF

Increased attention towards progress information near a goal state.

Psychon Bull Rev

January 2025

Department of Psychology, McGill University, 2001 Av. McGill College, Montréal, QC, H3A 1G1, Canada.

A growing body of evidence across psychology suggests that (cognitive) effort exertion increases in proximity to a goal state. For instance, previous work has shown that participants respond more quickly, but not less accurately, when they near a goal-as indicated by a filling progress bar. Yet it remains unclear when over the course of a cognitively demanding task do people monitor progress information: Do they continuously monitor their goal progress over the course of a task, or attend more frequently to it as they near their goal? To answer this question, we used eye-tracking to examine trial-by-trial changes in progress monitoring as participants completed blocks of an attentionally demanding oddball task.

View Article and Find Full Text PDF

How are arbitrary sequences of verbal information retained and manipulated in working memory? Increasing evidence suggests that serial order in verbal WM is spatially coded and that spatial attention is involved in access and retrieval. Based on the idea that brain areas controlling spatial attention are also involved in oculomotor control, we used eye tracking to reveal how the spatial structure of serial order information is accessed in verbal working memory. In two experiments, participants memorized a sequence of auditory words in the correct order.

View Article and Find Full Text PDF

Multi-item retro-cueing effects refer to better working memory performance for multiple items when they are cued after their offset compared to a neutral condition in which all items are cued. However, several studies have reported boundary conditions, and findings have also sometimes failed to replicate. We hypothesized that a strategy to focus on only one of the cued items could possibly yield these inconsistent patterns.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!