Attentional selection is a function that allocates the brain's computational resources to the most important part of a visual scene at a specific moment. Saliency map models have been proposed as computational models to predict attentional selection within a spatial location. Recent saliency map models based on deep convolutional neural networks (DCNNs) exhibit the highest performance for predicting the location of attentional selection and human gaze, which reflect overt attention. Trained DCNNs potentially provide insight into the perceptual mechanisms of biological visual systems. However, the relationship between artificial and neural representations used for determining attentional selection and gaze location remains unknown. To understand the mechanism underlying saliency map models based on DCNNs and the neural system of attentional selection, we investigated the correspondence between layers of a DCNN saliency map model and monkey visual areas for natural image representations. We compared the characteristics of the responses in each layer of the model with those of the neural representation in the primary visual (V1), intermediate visual (V4), and inferior temporal (IT) cortices. Regardless of the DCNN layer level, the characteristics of the responses were consistent with that of the neural representation in V1. We found marked peaks of correspondence between V1 and the early level and higher-intermediate-level layers of the model. These results provide insight into the mechanism of the trained DCNN saliency map model and suggest that the neural representations in V1 play an important role in computing the saliency that mediates attentional selection, which supports the V1 saliency hypothesis.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7890521PMC
http://dx.doi.org/10.1523/ENEURO.0200-20.2020DOI Listing

Publication Analysis

Top Keywords

saliency map
24
attentional selection
24
map model
12
map models
12
monkey visual
8
saliency
8
based deep
8
deep convolutional
8
convolutional neural
8
models based
8

Similar Publications

Chronic obstructive pulmonary disease (COPD) is a leading cause of death worldwide and greatly reduces the quality of life. Utilizing remote monitoring has been shown to improve quality of life and reduce exacerbations, but remains an ongoing area of research. We introduce a novel method for estimating changes in ease of breathing for COPD patients, using obstructed breathing data collected via wearables.

View Article and Find Full Text PDF

Rad4XCNN: A new agnostic method for post-hoc global explanation of CNN-derived features by means of Radiomics.

Comput Methods Programs Biomed

January 2025

Department of Biomedicine, Neuroscience and Advanced Diagnostics (BiND), University of Palermo, Palermo, 90127, Italy. Electronic address:

Article Synopsis
  • Machine learning-based clinical decision support systems (CDSS) face challenges with transparency and reliability, as explainability often reduces predictive accuracy.
  • A novel method called Rad4XCNN enhances the predictive power of CNN features while maintaining interpretability through Radiomics, moving beyond traditional saliency maps.
  • In breast cancer classification tasks, Rad4XCNN demonstrates superior accuracy compared to other feature types and allows for global insights, effectively addressing the explainability-accuracy trade-off in AI models.
View Article and Find Full Text PDF

A salience map is a topographic map that has inputs at each x,y location from many different feature maps and summarizes the combined salience of all those inputs as a real number, salience, which is represented in the map. Of the more than 1 million Google references to salience maps, nearly all use the map for computing the relative priority of visual image components for subsequent processing. We observe that salience processing is an instance of substance-invariant processing, analogous to household measuring cups, weight scales, and measuring tapes, all of which make single-number substance-invariant measurements.

View Article and Find Full Text PDF

Edge-guided feature fusion network for RGB-T salient object detection.

Front Neurorobot

December 2024

Department of Information Engineering, Shanghai Maritime University, Shanghai, China.

Introduction: RGB-T Salient Object Detection (SOD) aims to accurately segment salient regions in both visible light and thermal infrared images. However, many existing methods overlook the critical complementarity between these modalities, which can enhance detection accuracy.

Methods: We propose the Edge-Guided Feature Fusion Network (EGFF-Net), which consists of cross-modal feature extraction, edge-guided feature fusion, and salience map prediction.

View Article and Find Full Text PDF

The intelligent identification of wear particles in ferrography is a critical bottleneck that hampers the development and widespread adoption of ferrography technology. To address challenges such as false detection, missed detection of small wear particles, difficulty in distinguishing overlapping and similar abrasions, and handling complex image backgrounds, this paper proposes an algorithm called TCBGY-Net for detecting wear particles in ferrography images. The proposed TCBGY-Net uses YOLOv5s as the backbone network, which is enhanced with several advanced modules to improve detection performance.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!