Panoramic video provides immersive and interactive experience by enabling humans to control the field of view (FoV) through head movement (HM). Thus, HM plays a key role in modeling human attention on panoramic video. This paper establishes a database collecting subjects' HM in panoramic video sequences. From this database, we find that the HM data are highly consistent across subjects. Furthermore, we find that deep reinforcement learning (DRL) can be applied to predict HM positions, via maximizing the reward of imitating human HM scanpaths through the agent's actions. Based on our findings, we propose a DRL-based HM prediction (DHP) approach with offline and online versions, called offline-DHP and online-DHP. In offline-DHP, multiple DRL workflows are run to determine potential HM positions at each panoramic frame. Then, a heat map of the potential HM positions, named the HM map, is generated as the output of offline-DHP. In online-DHP, the next HM position of one subject is estimated given the currently observed HM position, which is achieved by developing a DRL algorithm upon the learned offline-DHP model. Finally, the experiments validate that our approach is effective in both offline and online prediction of HM positions for panoramic video, and that the learned offline-DHP model can improve the performance of online-DHP.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TPAMI.2018.2858783 | DOI Listing |
Video-assisted thoracic surgery (VATS) is a minimally invasive approach for treating early-stage non-small-cell lung cancer. Optimal trocar placement during VATS ensures comprehensive access to the thoracic cavity, provides a panoramic endoscopic view, and prevents instrument crowding. While established principles such as the Baseball Diamond Principle (BDP) and Triangle Target Principle (TTP) exist, surgeons mainly rely on experience and patient-specific anatomy for trocar placement, potentially leading to sub-optimal surgical plans that increase operative time and fatigue.
View Article and Find Full Text PDFFront Cell Dev Biol
October 2024
Ophthalmic Center, The Second Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang, China.
Aim: This study aimed to predict the formation of OBL during femtosecond laser SMILE surgery by employing deep learning technology.
Methods: This was a cross-sectional, retrospective study conducted at a university hospital. Surgical videos were randomly divided into a training (3,271 patches, 73.
Oral Maxillofac Surg
November 2024
Department of Oral and Maxillofacial Surgery, Kobe University Graduate School of Medicine, Kobe, Japan.
Purpose: This study was performed to investigate the association of dental dysplasia with childhood cancer. We examined the occurrence of agenesis, microdontia, and enamel changes of permanent teeth in pediatric cancer survivors.
Methods: Seventy-six patients with pediatric cancer and hematologic diseases were referred to our department for the first time from October 2005 to December 2019.
IEEE Trans Pattern Anal Mach Intell
October 2024
Event cameras are innovative neuromorphic sensors that asynchronously capture the scene dynamics. Due to the event-triggering mechanism, such cameras record event streams with much shorter response latency and higher intensity sensitivity compared to conventional cameras. On the basis of these features, previous works have attempted to reconstruct high dynamic range (HDR) videos from events, but have either suffered from unrealistic artifacts or failed to provide sufficiently high frame rates.
View Article and Find Full Text PDFLancet Infect Dis
September 2024
Nuffield Department of Primary Care Health Sciences, University of Oxford, Oxford, UK. Electronic address:
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!