Scene understanding is a key technical challenge within the autonomous driving domain. It requires a deep semantic understanding of the entities and relations found within complex physical and social environments that is both accurate and complete. In practice, this can be accomplished by representing entities in a scene and their relations as a knowledge graph (KG). This scene knowledge graph may then be utilized for the task of entity prediction, leading to improved scene understanding. In this paper, we will define and formalize this problem as Knowledge-based Entity Prediction (KEP). KEP aims to improve scene understanding by predicting potentially unrecognized entities by leveraging heterogeneous, high-level semantic knowledge of driving scenes. An innovative neuro-symbolic solution for KEP is presented, based on knowledge-infused learning, which 1) introduces a dataset agnostic ontology to describe driving scenes, 2) uses an expressive, holistic representation of scenes with knowledge graphs, and 3) proposes an effective, non-standard mapping of the KEP problem to the problem of link prediction (LP) using knowledge-graph embeddings (KGE). Using real, complex and high-quality data from urban driving scenes, we demonstrate its effectiveness by showing that the missing entities may be predicted with high precision (0.87 Hits@1) while significantly outperforming the non-semantic/rule-based baselines.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC8656233 | PMC |
http://dx.doi.org/10.3389/fdata.2021.759110 | DOI Listing |
J Neurophysiol
January 2025
Department of Sports Science, Zhejiang University, Hangzhou, Zhejiang, China.
Human postural control system has the capacity to adapt to balance-challenging perturbations. However, the characteristics and mechanisms of postural adaptation to continuous perturbation under the sensory conflicting environments remain unclear. We aimed to investigate the functional role of oscillatory coupling drive to lower-limb muscles with changes in balance control during postural adaptation under multisensory congruent and incongruent environments.
View Article and Find Full Text PDFPsychophysiology
January 2025
Department of Psychology, University of Georgia, Athens, Georgia, USA.
Emotional experiences involve dynamic multisensory perception, yet most EEG research uses unimodal stimuli such as naturalistic scene photographs. Recent research suggests that realistic emotional videos reliably reduce the amplitude of a steady-state visual evoked potential (ssVEP) elicited by a flickering border. Here, we examine the extent to which this video-ssVEP measure compares with the well-established Late Positive Potential (LPP) that is reliably larger for emotional relative to neutral scenes.
View Article and Find Full Text PDFAtten Percept Psychophys
January 2025
U.S. DEVCOM Army Research Laboratory, Humans in Complex Systems, Aberdeen Proving Ground, MD, USA.
Historically, electrophysiological correlates of scene processing have been studied with experiments using static stimuli presented for discrete timescales where participants maintain a fixed eye position. Gaps remain in generalizing these findings to real-world conditions where eye movements are made to select new visual information and where the environment remains stable but changes with our position and orientation in space, driving dynamic visual stimulation. Co-recording of eye movements and electroencephalography (EEG) is an approach to leverage fixations as time-locking events in the EEG recording under free-viewing conditions to create fixation-related potentials (FRPs), providing a neural snapshot in which to study visual processing under naturalistic conditions.
View Article and Find Full Text PDFMethodsX
June 2025
Department of Computer Engineering, Pimpri Chinchwad College of Engineering, Nigdi, Pune 411044, India.
Recent advancements in artificial intelligence (AI) have increased interest in intelligent transportation systems, particularly autonomous vehicles. Safe navigation in traffic-heavy environments requires accurate road scene segmentation, yet traditional computer vision methods struggle with complex scenarios. This study emphasizes the role of deep learning in improving semantic segmentation using datasets like the Indian Driving Dataset (IDD), which presents unique challenges in chaotic road conditions.
View Article and Find Full Text PDFSci Rep
January 2025
School of Information Engineering, Henan University of Science and Technology, Luoyang, 471000, China.
Lane detection is one of the key functions to ensure the safe driving of autonomous vehicles, and it is a challenging task. In real driving scenarios, external factors inevitably interfere with the lane detection system, such as missing lane markings, harsh weather conditions, and vehicle occlusion. To enhance the accuracy and detection speed of lane detection in complex road environments, this paper proposes an end-to-end lane detection model with a pure Transformer architecture, which exhibits excellent detection performance in complex road scenes.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!