Visual attention prediction (VAP) is a significant and imperative issue in the field of computer vision. Most of the existing VAP methods are based on deep learning. However, they do not fully take advantage of the low-level contrast features while generating the visual attention map. In this article, a novel VAP method is proposed to generate the visual attention map via bio-inspired representation learning. The bio-inspired representation learning combines both low-level contrast and high-level semantic features simultaneously, which are developed by the fact that the human eye is sensitive to the patches with high contrast and objects with high semantics. The proposed method is composed of three main steps: 1) feature extraction; 2) bio-inspired representation learning; and 3) visual attention map generation. First, the high-level semantic feature is extracted from the refined VGG16, while the low-level contrast feature is extracted by the proposed contrast feature extraction block in a deep network. Second, during bio-inspired representation learning, both the extracted low-level contrast and high-level semantic features are combined by the designed densely connected block, which is proposed to concatenate various features scale by scale. Finally, the weighted-fusion layer is exploited to generate the ultimate visual attention map based on the obtained representations after bio-inspired representation learning. Extensive experiments are performed to demonstrate the effectiveness of the proposed method.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TCYB.2019.2931735DOI Listing

Publication Analysis

Top Keywords

bio-inspired representation
24
representation learning
24
visual attention
24
low-level contrast
16
attention map
16
high-level semantic
12
learning visual
8
attention prediction
8
contrast high-level
8
semantic features
8

Similar Publications

Security is one of the increasingly significant issues given advancements in technology that harness data from multiple devices such as the internet of medical devices. While protecting data from unauthorized user access, several techniques are used including fingerprints, passwords, and others. One of the techniques that has attracted much attention is the use of human features, which has proven to be most effective because of the difficulties in impersonating human-related features.

View Article and Find Full Text PDF
Article Synopsis
  • Event cameras mimic biological vision, providing high dynamic range and temporal resolution, which improve motion estimation in complex scenes, but they have limitations like low spatial resolution and data redundancy.
  • Current pre-processing methods for event streams struggle with challenges such as varying speeds and multiple subjects, leading to inefficiencies in data handling.
  • The proposed Asynchronous Spike Dynamic Metric and Slicing algorithm (ASDMS) and Adaptive Spatiotemporal Subject Surface Compensation algorithm (ASSSC) enhance event stream segmentation and image quality, while a new evaluation metric, Actual Performance Efficiency Discrepancy (APED), measures the effectiveness of these improvements against existing methods.
View Article and Find Full Text PDF

Multi-compartment neuron and population encoding powered spiking neural network for deep distributional reinforcement learning.

Neural Netw

February 2025

Brain-inspired Cognitive Intelligence Lab, Institute of Automation, Chinese Academy of Sciences, Beijing, 100190, China; Center for Long-term Artificial Intelligence, Beijing, 100190, China; University of Chinese Academy of Sciences, Beijing, 100049, China; Key Laboratory of Brain Cognition and Brain-inspired Intelligence Technology, Chinese Academy of Sciences, Shanghai, 200031, China. Electronic address:

Inspired by the brain's information processing using binary spikes, spiking neural networks (SNNs) offer significant reductions in energy consumption and are more adept at incorporating multi-scale biological characteristics. In SNNs, spiking neurons serve as the fundamental information processing units. However, in most models, these neurons are typically simplified, focusing primarily on the leaky integrate-and-fire (LIF) point neuron model while neglecting the structural properties of biological neurons.

View Article and Find Full Text PDF
Article Synopsis
  • * The researchers used a signal extraction method to analyze the clicks in terms of their timing and frequency, which allows them to visualize energy distribution in a 2D format.
  • * The findings suggest that this bio-inspired click train can enhance sonar capabilities, enabling better target discrimination by adjusting energy focus based on the properties of the target, indicating potential for advanced man-made sonar applications.
View Article and Find Full Text PDF
Article Synopsis
  • - The study focuses on improving blurry image deblurring by using event cameras, which are bio-inspired visual sensors, to recover sharp frames from blurry images.
  • - It critiques existing methods that overlook the importance of blur modeling and proposes a new approach that uses an event-assisted blurriness encoder to implicitly model blur in images.
  • - The method integrates blurriness information into a base image unfolding network using novel modulation and aggregation techniques, demonstrating superior performance in experiments compared to current state-of-the-art methods on various datasets.
View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!