Learning sparse and meaningful representations through embodiment.

Neural Netw

Institute of Cognitive Science, University of Osnabrück, Wachsbleiche 27, 49090 Osnabrück, Germany.

Published: February 2021

How do humans acquire a meaningful understanding of the world with little to no supervision or semantic labels provided by the environment? Here we investigate embodiment with a closed loop between action and perception as one key component in this process. We take a close look at the representations learned by a deep reinforcement learning agent that is trained with high-dimensional visual observations collected in a 3D environment with very sparse rewards. We show that this agent learns stable representations of meaningful concepts such as doors without receiving any semantic labels. Our results show that the agent learns to represent the action relevant information, extracted from a simulated camera stream, in a wide variety of sparse activation patterns. The quality of the representations learned shows the strength of embodied learning and its advantages over fully supervised approaches.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.neunet.2020.11.004DOI Listing

Publication Analysis

Top Keywords

semantic labels
8
representations learned
8
agent learns
8
learning sparse
4
sparse meaningful
4
representations
4
meaningful representations
4
representations embodiment
4
embodiment humans
4
humans acquire
4

Similar Publications

Human activity recognition by radar sensors plays an important role in healthcare and smart homes. However, labeling a large number of radar datasets is difficult and time-consuming, and it is difficult for models trained on insufficient labeled data to obtain exact classification results. In this paper, we propose a multiscale residual weighted classification network with large-scale, medium-scale, and small-scale residual networks.

View Article and Find Full Text PDF

Data-Efficient Bone Segmentation Using Feature Pyramid- Based SegFormer.

Sensors (Basel)

December 2024

Master's Program in Information and Computer Science, Doshisha University, Kyoto 610-0394, Japan.

The semantic segmentation of bone structures demands pixel-level classification accuracy to create reliable bone models for diagnosis. While Convolutional Neural Networks (CNNs) are commonly used for segmentation, they often struggle with complex shapes due to their focus on texture features and limited ability to incorporate positional information. As orthopedic surgery increasingly requires precise automatic diagnosis, we explored SegFormer, an enhanced Vision Transformer model that better handles spatial awareness in segmentation tasks.

View Article and Find Full Text PDF

Visual semantic decoding aims to extract perceived semantic information from the visual responses of the human brain and convert it into interpretable semantic labels. Although significant progress has been made in semantic decoding across individual visual cortices, studies on the semantic decoding of the ventral and dorsal cortical visual pathways remain limited. This study proposed a graph neural network (GNN)-based semantic decoding model on a natural scene dataset (NSD) to investigate the decoding differences between the dorsal and ventral pathways in process various parts of speech, including verbs, nouns, and adjectives.

View Article and Find Full Text PDF

Enhanced diagnosis of pes planus and pes cavus using deep learning-based segmentation of weight-bearing lateral foot radiographs: a comparative observer study.

Biomed Eng Lett

January 2025

Department of Biomedical Engineering, Asan Medical Institute of Convergence Science and Technology, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Republic of Korea.

Unlabelled: A weight-bearing lateral radiograph (WBLR) of the foot is a gold standard for diagnosing adult-acquired flatfoot deformity. However, it is difficult to measure the major axis of bones in WBLR without using auxiliary lines. Herein, we develop semantic segmentation with a deep learning model (DLm) on the WBLR of the foot for enhanced diagnosis of pes planus and pes cavus.

View Article and Find Full Text PDF

Expanding the concept of ID conversion in TogoID by introducing multi-semantic and label features.

J Biomed Semantics

January 2025

Database Center for Life Science, Joint Support-Center for Data Science Research, Research Organization of Information and Systems, Kashiwa, Chiba, Japan.

Background: TogoID ( https://togoid.dbcls.jp/ ) is an identifier (ID) conversion service designed to link IDs across diverse categories of life science databases.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!