Neural surprise in somatosensory Bayesian learning.

PLoS Comput Biol

Neurocomputation and Neuroimaging Unit, Freie Universität Berlin, Germany.

Published: February 2021

Tracking statistical regularities of the environment is important for shaping human behavior and perception. Evidence suggests that the brain learns environmental dependencies using Bayesian principles. However, much remains unknown about the employed algorithms, for somesthesis in particular. Here, we describe the cortical dynamics of the somatosensory learning system to investigate both the form of the generative model as well as its neural surprise signatures. Specifically, we recorded EEG data from 40 participants subjected to a somatosensory roving-stimulus paradigm and performed single-trial modeling across peri-stimulus time in both sensor and source space. Our Bayesian model selection procedure indicates that evoked potentials are best described by a non-hierarchical learning model that tracks transitions between observations using leaky integration. From around 70ms post-stimulus onset, secondary somatosensory cortices are found to represent confidence-corrected surprise as a measure of model inadequacy. Indications of Bayesian surprise encoding, reflecting model updating, are found in primary somatosensory cortex from around 140ms. This dissociation is compatible with the idea that early surprise signals may control subsequent model update rates. In sum, our findings support the hypothesis that early somatosensory processing reflects Bayesian perceptual learning and contribute to an understanding of its underlying mechanisms.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7880500PMC
http://dx.doi.org/10.1371/journal.pcbi.1008068DOI Listing

Publication Analysis

Top Keywords

neural surprise
8
somatosensory
6
model
6
bayesian
5
surprise somatosensory
4
somatosensory bayesian
4
learning
4
bayesian learning
4
learning tracking
4
tracking statistical
4

Similar Publications

Facial expression recognition (FER) has caught the research community's attention recently because it can affect many real-life applications. Multiple studies have focused on automatic FER, most of which use a machine learning methodology, FER has continued to be a difficult and exciting issue in computer vision. Deep learning has recently drawn increased attention as a solution to several practical issues, including facial expression recognition.

View Article and Find Full Text PDF
Article Synopsis
  • Humans have a strong ability to remember scene images, retaining more detail than previously thought, allowing them to differentiate between similar scenes.
  • Research shows that the complexity of a scene can positively correlate with how memorable it is, challenging established theories about this relationship.
  • Studies conducted reveal that while complexity and memorability are interconnected, their relationship is not straightforward and can be influenced by cognitive load, further suggesting complexity may serve as a useful measure for scene detail in memory studies.
View Article and Find Full Text PDF

Facial expression analysis using convolutional neural network for drug-naive and chronic schizophrenia.

J Psychiatr Res

November 2024

Institute of Mental Health, Tianjin Anding Hospital, Mental Health Center of Tianjin Medical University, Tianjin, China; Brain Assessment & Intervention Laboratory, Tianjin Anding Hospital, Mental Health Center of Tianjin Medical University, Tianjin, China. Electronic address:

Objective: Facial images have been shown to convey mental conditions as clinical symptoms. This study aimed to use facial images to detect patients with drug-naive schizophrenia (DN-SCZ) or chronic schizophrenia (C-SCZ) from healthy controls (HCs), and to investigate differences in facial expressions among these 3 groups, as well as relationships between facial expressions and psychiatric symptoms.

Methods: We recruited 45 DN-SCZ patients, 106 C-SCZ patients and 101 HCs for the study, and videotaped their facial expressions through a fixed experimental paradigm.

View Article and Find Full Text PDF

Assessing the effects of 5-HT and 5-HT receptor antagonists on DOI-induced head-twitch response in male rats using marker-less deep learning algorithms.

Pharmacol Rep

November 2024

Behavioral Neuroscience and Drug Development, Maj Institute of Pharmacology, Polish Academy of Sciences, Smętna 12, Kraków, 31-343, Poland.

Article Synopsis
  • The study investigates a new, marker-less method using deep learning to track head-twitch responses in rodents, traditionally observed by humans.
  • High-speed videos were analyzed with DeepLabCut and SimBA, showing strong agreement with human counts while evaluating the effects of the psychedelic DOI.
  • Results confirmed that certain 5-HT receptor antagonists can reduce head-twitch responses, supporting the idea that this behavior is specific to serotonergic activity, and demonstrated the effectiveness of the automated tracking tools for research purposes.
View Article and Find Full Text PDF

Neural Mechanism of Musical Pleasure Induced by Prediction Errors: An EEG Study.

Brain Sci

November 2024

Department of Electronics and Bioinformatics, School of Science and Technology, Meiji University, Kawasaki 214-8571, Japan.

Background/objectives: Musical pleasure is considered to be induced by prediction errors (surprise), as suggested in neuroimaging studies. However, the role of temporal changes in musical features in reward processing remains unclear. Utilizing the Information Dynamics of Music (IDyOM) model, a statistical model that calculates musical surprise based on prediction errors in melody and harmony, we investigated whether brain activities associated with musical pleasure, particularly in the θ, β, and γ bands, are induced by prediction errors, similar to those observed during monetary rewards.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!