Appearance-based gaze estimation has garnered increasing attention in recent years. However, deep learning-based gaze estimation models still suffer from suboptimal performance when deployed in new domains, e.g., unseen environments or individuals. In our previous work, we took this challenge for the first time by introducing a plug-and-play method (PnP-GA) to adapt the gaze estimation model to new domains. The core concept of PnP-GA is to leverage the diversity brought by a group of model variants to enhance the adaptability to diverse environments. In this article, we propose the PnP-GA+ by extending our approach to explore the impact of assembling model variants using three additional perspectives: color space, data augmentation, and model structure. Moreover, we propose an intra-group attention module that dynamically optimizes pseudo-labeling during adaptation. Experimental results demonstrate that by directly plugging several existing gaze estimation networks into the PnP-GA+ framework, it outperforms state-of-the-art domain adaptation approaches on four standard gaze domain adaptation tasks on public datasets. Our method consistently enhances cross-domain performance, and its versatility is improved through various ways of assembling the model group.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TPAMI.2023.3348528 | DOI Listing |
eNeuro
January 2025
Department of Rehabilitation Medicine, Kansai Medical University, 2-5-1 Shin-machi, Hirakata, Osaka 573-1010, Japan.
The subjective visual vertical (VV), the visually estimated direction of gravity, is essential for assessing vestibular function and visuospatial cognition. In this study, we aimed to investigate the mechanisms underlying altered VV perception in stroke participants with unilateral spatial neglect (USN), specifically by examining their eye movement patterns during VV judgment tasks. Participants with USN demonstrated limited eye movement scanning along a rotating bar, often fixating on prominent ends, such as the top or bottom.
View Article and Find Full Text PDFJ Vis
January 2025
Department of Cognitive and Psychological Sciences, Graduate School of Informatics, Nagoya University, Aichi, Japan.
Humans can estimate the time and position of a moving object's arrival. However, numerous studies have demonstrated superior position estimation accuracy for descending objects compared with ascending objects. We tested whether the accuracy of position estimation for ascending and descending objects differs between the upper and lower visual fields.
View Article and Find Full Text PDFCureus
November 2024
Anesthesiology and Pain Medicine, Harborview Medical Center, Seattle, USA.
Behav Res Methods
December 2024
Lund University Humanities Lab, Lund University, Lund, Sweden.
Irrespective of the precision, the inaccuracy of a pupil-based eye tracker is about 0.5 . This paper delves into two factors that potentially increase the inaccuracy of the gaze signal, namely, 1) Pupil-size changes and the pupil-size artefact (PSA) and 2) the putative inability of experienced individuals to precisely refixate a visual target.
View Article and Find Full Text PDFCurr Biol
December 2024
Behavioral and Systems Neuroscience, Department of Psychology, Rutgers University, New Brunswick, NJ 08854, USA. Electronic address:
Determining the location of objects relative to ourselves is essential for interacting with the world. Neural activity in the retina is used to form a vision-independent model of the local spatial environment relative to the body. For example, when an animal navigates through a forest, it rapidly shifts its gaze to identify the position of important objects, such as a tree obstructing its path.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!