Large Language Models (LLMs), such as the General Pre-trained Transformer (GPT), have shown remarkable performance in various cognitive tasks. However, it remains unclear whether these models have the ability to accurately infer human perceptual representations. Previous research has addressed this question by quantifying correlations between similarity response patterns of humans and LLMs. Correlation provides a measure of similarity, but it relies pre-defined item labels and does not distinguish category- and item- level similarity, falling short of characterizing detailed structural correspondence between humans and LLMs. To assess their structural equivalence in more detail, we propose the use of an unsupervised alignment method based on Gromov-Wasserstein optimal transport (GWOT). GWOT allows for the comparison of similarity structures without relying on pre-defined label correspondences and can reveal fine-grained structural similarities and differences that may not be detected by simple correlation analysis. Using a large dataset of similarity judgments of 93 colors, we compared the color similarity structures of humans (color-neurotypical and color-atypical participants) and two GPT models (GPT-3.5 and GPT-4). Our results show that the similarity structure of color-neurotypical participants can be remarkably well aligned with that of GPT-4 and, to a lesser extent, to that of GPT-3.5. These results contribute to the methodological advancements of comparing LLMs with human perception, and highlight the potential of unsupervised alignment methods to reveal detailed structural correspondences.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11237038 | PMC |
http://dx.doi.org/10.1038/s41598-024-65604-1 | DOI Listing |
Front Oncol
January 2025
Department of Radiation Oncology, Yonsei Cancer Center, Heavy Ion Therapy Research Institute, Yonsei University College of Medicine, Seoul, Republic of Korea.
Purpose: Recent deep-learning based synthetic computed tomography (sCT) generation using magnetic resonance (MR) images have shown promising results. However, generating sCT for the abdominal region poses challenges due to the patient motion, including respiration and peristalsis. To address these challenges, this study investigated an unsupervised learning approach using a transformer-based cycle-GAN with structure-preserving loss for abdominal cancer patients.
View Article and Find Full Text PDFNeural Netw
January 2025
School of Computer Engineering and Science, Shanghai University, Shanghai, 200444, China. Electronic address:
Conditional adversarial domain adaptation (CADA) is one of the most commonly used unsupervised domain adaptation (UDA) methods. CADA introduces multimodal information to the adversarial learning process to align the distributions of the labeled source domain and unlabeled target domain with mode match. However, CADA provides wrong multimodal information for challenging target features due to utilizing classifier predictions as the multimodal information, leading to distribution mismatch and less robust domain-invariant features.
View Article and Find Full Text PDFPhys Med Biol
January 2025
School of Software Engineering, Xi'an Jiaotong University, Xi 'an Jiaotong University Innovation Port, Xi 'an, Shaanxi Province, Xi'an, Shaanxi, 710049, CHINA.
Deformable registration aims to achieve nonlinear alignment of image space by estimating a dense displacement field. It is commonly used as a preprocessing step in clinical and image analysis applications, such as surgical planning, diagnostic assistance, and surgical navigation. We aim to overcome these challenges: Deep learning-based registration methods often struggle with complex displacements and lack effective interaction between global and local feature information.
View Article and Find Full Text PDFSensors (Basel)
January 2025
School of Mechanical and Electrical Engineering, China University of Mining and Technology (Beijing), Beijing 100083, China.
Unsupervised Domain Adaptation for Object Detection (UDA-OD) aims to adapt a model trained on a labeled source domain to an unlabeled target domain, addressing challenges posed by domain shifts. However, existing methods often face significant challenges, particularly in detecting small objects and over-relying on classification confidence for pseudo-label selection, which often leads to inaccurate bounding box localization. To address these issues, we propose a novel UDA-OD framework that leverages scale consistency (SC) and Temporal Ensemble Pseudo-Label Selection (TEPLS) to enhance cross-domain robustness and detection performance.
View Article and Find Full Text PDFNat Commun
January 2025
Bernstein Center for Computational Neuroscience Berlin, Humboldt-Universität zu Berlin, Berlin, Germany.
Understanding vibrissal transduction has advanced by serial sectioning and identified afferent recordings, but afferent mapping onto the complex, encapsulated follicle remains unclear. Here, we reveal male rat C2 vibrissa follicle innervation through synchrotron X-ray phase contrast tomograms. Morphological analysis identified 5% superficial, ~32 % unmyelinated and 63% myelinated deep vibrissal nerve axons.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!