A number of recent models of semantics combine linguistic information, derived from text corpora, and visual information, derived from image collections, demonstrating that the resulting multimodal models are better than either of their unimodal counterparts, in accounting for behavioral data. Empirical work on semantic processing has shown that emotion also plays an important role especially in abstract concepts; however, models integrating emotion along with linguistic and visual information are lacking. Here, we first improve on visual and affective representations, derived from state-of-the-art existing models, by choosing models that best fit available human semantic data and extending the number of concepts they cover. Crucially then, we assess whether adding affective representations (obtained from a neural network model designed to predict emojis from co-occurring text) improves the model's ability to fit semantic similarity/relatedness judgments from a purely linguistic and linguistic-visual model. We find that, given specific weights assigned to the models, adding both visual and affective representations improves performance, with visual representations providing an improvement especially for more concrete words, and affective representations improving especially the fit for more abstract words.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1111/cogs.12830 | DOI Listing |
Proc Natl Acad Sci U S A
January 2025
Department of Psychology, City College, City University of New York, New York, NY 10031.
Looking at the world often involves not just seeing things, but feeling things. Modern feedforward machine vision systems that learn to perceive the world in the absence of active physiology, deliberative thought, or any form of feedback that resembles human affective experience offer tools to demystify the relationship between seeing and feeling, and to assess how much of visually evoked affective experiences may be a straightforward function of representation learning over natural image statistics. In this work, we deploy a diverse sample of 180 state-of-the-art deep neural network models trained only on canonical computer vision tasks to predict human ratings of arousal, valence, and beauty for images from multiple categories (objects, faces, landscapes, art) across two datasets.
View Article and Find Full Text PDFInfant Ment Health J
January 2025
Crown Family School of Social Work, Policy, and Practice, The University of Chicago, Chicago, Illinois, USA.
Although mother-to-infant attachment begins during pregnancy, few studies have explored correlates of prenatal attachment and associations with later measures of attachment representations. This study explored whether prenatal attachment is related to attachment representations during toddlerhood and whether associations between them reflect the broader quality of mothers' relationships. Young, ethnically/racially diverse, low-income American women (n = 160) were followed from pregnancy through 30 months postpartum.
View Article and Find Full Text PDFCogn Emot
January 2025
Brain and Cognition, KU Leuven, Leuven, Belgium.
Decision confidence is a prototypical metacognitive representation that is thought to approximate the probability that a decision is correct. The perception of being correct has also been associated with affective valence such that being correct feels more positive and being mistaken more negative. This suggests that, similarly to confidence, affective valence reflects the probability that a decision is correct.
View Article and Find Full Text PDFJ Cogn
January 2025
General Psychology, Trier University, Germany.
Observations from multisensory body illusions indicate that the body representation can be adapted to changing task demands, e.g., it can be expanded to integrate external objects based on current sensorimotor experience (embodiment).
View Article and Find Full Text PDFBrain Lang
January 2025
Univ. Lille, CNRS, UMR 9193 - SCALab - Sciences Cognitives et Sciences Affectives, F-59000 Lille, France; Univ. Lille, Inria, CNRS, Centrale Lille, UMR 9189 - CRIStAL, F-59000, Lille, France. Electronic address:
Although previous research has shown that speakers adapt on the words they use, it remains unclear whether speakers adapt their phonological representations, leading them to perceive new phonemic contrasts following a social interaction. This event-related potential (ERP) study investigates whether the neuronal responses to the perception of the /e/-/ε/ vowel merger in Northern French speakers show evidence for discriminating /e/ and /ε/ phonemes after interacting with a speaker who produced this contrast. Northern French participants engaged in an interactive map task and we measured their ERP responses elicited after the presentation of a last syllable which was either phonemically identical to or different from preceding syllables.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!