The current study investigates whether informative, mutually redundant audiovisual cues support better performance in a category learning paradigm. Research suggests that, under some conditions, redundant multisensory cues supports better learning, when compared with unisensory cues. This was examined systematically across two experiments. In Experiment 1, children aged 5, 7, and 10 years were allocated to 1 of the 3 "modality" conditions (audio informative only, visual informative only, and audiovisual informative) and explicitly instructed to learn the category membership of individual exemplars, as determined by a threshold of correct responses. Unisensory or redundant multisensory cues determined category membership, depending on the learning condition. In addition to significant main effects of age group and condition, a significant interaction between age and sensory condition was found, with 5-year-olds performing better when presented with redundant multisensory cues compared to unisensory cues. Ten-year-olds performed better with auditory informative only cues, compared to visual informative only cues, or informative but redundant multisensory cues, with no significant difference between the latter two. In Experiment 2, the multisensory condition was presented to separate groups of 5-, 7-, and 10-year-olds, examining explicit learning outcomes in the audiovisual informative condition. Results showed that children who reached threshold during training were faster, made fewer errors, and performed better during test trials. Learning appeared to be based on the visual informative cues. Findings are discussed in the context of age-related selective attention, suggesting that the value of providing multisensory informative cues to support real-world learning depends on age and instructional context. (PsycINFO Database Record (c) 2019 APA, all rights reserved).
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1037/dev0000778 | DOI Listing |
J Neurosci
November 2024
Department of Speech, Language, and Hearing Sciences, Moody College of Communication, The University of Texas at Austin, Austin, Texas 78712
The human auditory cortex is organized according to the timing and spectral characteristics of speech sounds during speech perception. During listening, the posterior superior temporal gyrus is organized according to onset responses, which segment acoustic boundaries in speech, and sustained responses, which further process phonological content. When we speak, the auditory system is actively processing the sound of our own voice to detect and correct speech errors in real time.
View Article and Find Full Text PDFbioRxiv
May 2024
Department of Speech, Language, and Hearing Sciences, Moody College of Communication, The University of Texas at Austin, Austin, TX, USA.
Psychon Bull Rev
October 2024
Institute of Cognitive Sciences and Technologies, National Research Council of Italy (CNR), Rome, Italy.
The term 'amodal' is a key topic in several different research fields across experimental psychology and cognitive neuroscience, including in the areas of developmental and perception science. However, despite being regularly used in the literature, the term means something different to the researchers working in the different contexts. Many developmental scientists conceive of the term as referring to those perceptual qualities, such as, for example, the size and shape of an object, that can be picked up by multiple senses (e.
View Article and Find Full Text PDFSci Rep
January 2024
School of Psychology and Neuroscience, University of St. Andrews, St. Mary's Quad, South Street, St. Andrews, KY16 9JP, UK.
Responses to multisensory signals are often faster compared to their unisensory components. This speed-up is typically attributed to target redundancy in that a correct response can be triggered by one or the other signal. In addition, semantic congruency of signals can also modulate multisensory responses; however, the contribution of semantic content is difficult to isolate as its manipulation commonly changes signal redundancy as well.
View Article and Find Full Text PDFFront Comput Neurosci
October 2023
Department of Electrical, Electronic, and Information Engineering Guglielmo Marconi, University of Bologna, Bologna, Italy.
In everyday life, the brain processes a multitude of stimuli from the surrounding environment, requiring the integration of information from different sensory modalities to form a coherent perception. This process, known as multisensory integration, enhances the brain's response to redundant congruent sensory cues. However, it is equally important for the brain to segregate sensory inputs from distinct events, to interact with and correctly perceive the multisensory environment.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!