Interactively learning behavior trees from imperfect human demonstrations.

Front Robot AI

Interactive AI & Cognitive Models for Human-AI Interaction (IKIDA), Technische Universität Darmstadt, Darmstadt, Germany.

Published: July 2023

AI Article Synopsis

  • In Interactive Task Learning (ITL), an agent learns tasks through natural interactions with human instructors, but learning methods often require detailed step-by-step instructions or do not allow for easy modifications of learned tasks.
  • The proposed framework utilizes Behavior Trees (BTs) to learn tasks from just a few RGB-D video demonstrations by automatically extracting conditions for actions and allowing for real-time adjustments based on user interactions.
  • A user study with 20 participants focused on a robotic trash disposal task showed that this new method can efficiently learn and refine BTs in response to detected failures during task execution.

Article Abstract

In Interactive Task Learning (ITL), an agent learns a new task through natural interaction with a human instructor. Behavior Trees (BTs) offer a reactive, modular, and interpretable way of encoding task descriptions but have not yet been applied a lot in robotic ITL settings. Most existing approaches that learn a BT from human demonstrations require the user to specify each action step-by-step or do not allow for adapting a learned BT without the need to repeat the entire teaching process from scratch. We propose a new framework to directly learn a BT from only a few human task demonstrations recorded as RGB-D video streams. We automatically extract continuous pre- and post-conditions for BT action nodes from visual features and use a Backchaining approach to build a reactive BT. In a user study on how non-experts provide and vary demonstrations, we identify three common failure cases of an BT learned from potentially imperfect initial human demonstrations. We offer a way to interactively resolve these failure cases by refining the existing BT through interaction with a user over a web-interface. Specifically, failure cases or unknown states are detected automatically during the execution of a learned BT and the initial BT is adjusted or extended according to the provided user input. We evaluate our approach on a robotic trash disposal task with 20 human participants and demonstrate that our method is capable of learning reactive BTs from only a few human demonstrations and interactively resolving possible failure cases at runtime.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC10368948PMC
http://dx.doi.org/10.3389/frobt.2023.1152595DOI Listing

Publication Analysis

Top Keywords

human demonstrations
16
failure cases
16
behavior trees
8
learn human
8
human
7
demonstrations
6
task
5
interactively learning
4
learning behavior
4
trees imperfect
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!