Deep language models (DLMs) have exhibited remarkable language understanding and generation capabilities, prompting researchers to explore the similarities between their internal mechanisms and human language cognitive processing. This study investigated the representational similarity (RS) between the abstractive summarization (ABS) models and the human brain and its correlation to the performance of ABS tasks. Specifically, representational similarity analysis (RSA) was used to measure the similarity between the representational patterns (RPs) of the BART, PEGASUS, and T5 models' hidden layers and the human brain's language RPs under different spatiotemporal conditions. Layer-wise ablation manipulation, including attention ablation and noise addition was employed to examine the hidden layers' effect on model performance. The results demonstrate that as the depth of hidden layers increases, the models' text encoding becomes increasingly similar to the human brain's language RPs. Manipulating deeper layers leads to more substantial decline in summarization performance compared to shallower layers, highlighting the crucial role of deeper layers in integrating essential information. Notably, the study confirms the hypothesis that the hidden layers exhibiting higher similarity to human brain activity play a more critical role in model performance, with their correlations reaching statistical significance even after controlling for perplexity. These findings deepen our understanding of the cognitive mechanisms underlying language representations in DLMs and their neural correlates, potentially providing insights for optimizing and improving language models by aligning them with the human brain's language-processing mechanisms.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11696092PMC
http://dx.doi.org/10.1038/s41598-024-84530-wDOI Listing

Publication Analysis

Top Keywords

hidden layers
12
human brain's
12
abstractive summarization
8
summarization performance
8
language models
8
representational similarity
8
human brain
8
brain's language
8
language rps
8
model performance
8

Similar Publications

Energy efficiency in computation is ultimately limited by noise, with quantum limits setting the fundamental noise floor. Analog physical neural networks hold promise for improved energy efficiency compared to digital electronic neural networks. However, they are typically operated in a relatively high-power regime so that the signal-to-noise ratio (SNR) is large (>10), and the noise can be treated as a perturbation.

View Article and Find Full Text PDF

Fusion transcripts in plants: hidden layer of transcriptome complexity.

Trends Plant Sci

January 2025

Bioinformatics Laboratory, National Institute of Plant Genome Research, Aruna Asaf Ali Marg, New Delhi 110067, India. Electronic address:

In the realm of genetic information, fusion transcripts contribute to the intricate complexity of the transcriptome across various organisms. Recently, Cong et al. investigated these RNAs in rice, maize, soybean, and arabidopsis (Arabidopsis thaliana), revealing conserved characteristics.

View Article and Find Full Text PDF

Deep language models (DLMs) have exhibited remarkable language understanding and generation capabilities, prompting researchers to explore the similarities between their internal mechanisms and human language cognitive processing. This study investigated the representational similarity (RS) between the abstractive summarization (ABS) models and the human brain and its correlation to the performance of ABS tasks. Specifically, representational similarity analysis (RSA) was used to measure the similarity between the representational patterns (RPs) of the BART, PEGASUS, and T5 models' hidden layers and the human brain's language RPs under different spatiotemporal conditions.

View Article and Find Full Text PDF

Point mutations at codon 600 of the BRAF oncogene are the most common alterations in cutaneous melanoma (CM). Assessment of BRAF status allows to personalize patient management, though the affordability of molecular testing is limited in some countries. This study aimed to develop a model for predicting alteration in BRAF based on routinely available clinical and histological data.

View Article and Find Full Text PDF

This paper presents a ground motion prediction (GMP) model using an artificial neural network (ANN) for shallow earthquakes, aimed at improving earthquake hazard safety evaluation. The proposed model leverages essential input variables such as moment magnitude, fault type, epicentral distance, and soil type, with the output variable being peak ground acceleration (PGA) at 5% damping. To develop this model, 885 data pairs were obtained from the Pacific Engineering Research Center, providing a robust dataset for training and validation.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!