A number of recent models of semantics combine linguistic information, derived from text corpora, and visual information, derived from image collections, demonstrating that the resulting multimodal models are better than either of their unimodal counterparts, in accounting for behavioral data. Empirical work on semantic processing has shown that emotion also plays an important role especially in abstract concepts; however, models integrating emotion along with linguistic and visual information are lacking. Here, we first improve on visual and affective representations, derived from state-of-the-art existing models, by choosing models that best fit available human semantic data and extending the number of concepts they cover. Crucially then, we assess whether adding affective representations (obtained from a neural network model designed to predict emojis from co-occurring text) improves the model's ability to fit semantic similarity/relatedness judgments from a purely linguistic and linguistic-visual model. We find that, given specific weights assigned to the models, adding both visual and affective representations improves performance, with visual representations providing an improvement especially for more concrete words, and affective representations improving especially the fit for more abstract words.

Download full-text PDF

Source
http://dx.doi.org/10.1111/cogs.12830DOI Listing

Publication Analysis

Top Keywords

affective representations
16
visual affective
8
models
7
visual
5
representations
5
constructing semantic
4
semantic models
4
models images
4
images emojis
4
emojis number
4

Similar Publications

Looking at the world often involves not just seeing things, but feeling things. Modern feedforward machine vision systems that learn to perceive the world in the absence of active physiology, deliberative thought, or any form of feedback that resembles human affective experience offer tools to demystify the relationship between seeing and feeling, and to assess how much of visually evoked affective experiences may be a straightforward function of representation learning over natural image statistics. In this work, we deploy a diverse sample of 180 state-of-the-art deep neural network models trained only on canonical computer vision tasks to predict human ratings of arousal, valence, and beauty for images from multiple categories (objects, faces, landscapes, art) across two datasets.

View Article and Find Full Text PDF

Young mothers' prenatal attachment and later attachment-related representations of their young children.

Infant Ment Health J

January 2025

Crown Family School of Social Work, Policy, and Practice, The University of Chicago, Chicago, Illinois, USA.

Although mother-to-infant attachment begins during pregnancy, few studies have explored correlates of prenatal attachment and associations with later measures of attachment representations. This study explored whether prenatal attachment is related to attachment representations during toddlerhood and whether associations between them reflect the broader quality of mothers' relationships. Young, ethnically/racially diverse, low-income American women (n = 160) were followed from pregnancy through 30 months postpartum.

View Article and Find Full Text PDF

Metacognitive confidence and affect - two sides of the same coin?

Cogn Emot

January 2025

Brain and Cognition, KU Leuven, Leuven, Belgium.

Decision confidence is a prototypical metacognitive representation that is thought to approximate the probability that a decision is correct. The perception of being correct has also been associated with affective valence such that being correct feels more positive and being mistaken more negative. This suggests that, similarly to confidence, affective valence reflects the probability that a decision is correct.

View Article and Find Full Text PDF

Observations from multisensory body illusions indicate that the body representation can be adapted to changing task demands, e.g., it can be expanded to integrate external objects based on current sensorimotor experience (embodiment).

View Article and Find Full Text PDF

Neural evidence for perceiving a vowel merger after a social interaction within a native language.

Brain Lang

January 2025

Univ. Lille, CNRS, UMR 9193 - SCALab - Sciences Cognitives et Sciences Affectives, F-59000 Lille, France; Univ. Lille, Inria, CNRS, Centrale Lille, UMR 9189 - CRIStAL, F-59000, Lille, France. Electronic address:

Although previous research has shown that speakers adapt on the words they use, it remains unclear whether speakers adapt their phonological representations, leading them to perceive new phonemic contrasts following a social interaction. This event-related potential (ERP) study investigates whether the neuronal responses to the perception of the /e/-/ε/ vowel merger in Northern French speakers show evidence for discriminating /e/ and /ε/ phonemes after interacting with a speaker who produced this contrast. Northern French participants engaged in an interactive map task and we measured their ERP responses elicited after the presentation of a last syllable which was either phonemically identical to or different from preceding syllables.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!