Learning Human-to-Robot Handovers from Point Clouds
We propose the first framework to learn control policies for vision-based human-to-robot handovers, a critical task for human-robot interaction. While research in Embodied AI has made significant progress in training robot agents in simulated environments, interacting with humans remains challenging...
Saved in:
Main Authors: | , , , , , |
---|---|
Format: | Journal Article |
Language: | English |
Published: |
30-03-2023
|
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | We propose the first framework to learn control policies for vision-based
human-to-robot handovers, a critical task for human-robot interaction. While
research in Embodied AI has made significant progress in training robot agents
in simulated environments, interacting with humans remains challenging due to
the difficulties of simulating humans. Fortunately, recent research has
developed realistic simulated environments for human-to-robot handovers.
Leveraging this result, we introduce a method that is trained with a
human-in-the-loop via a two-stage teacher-student framework that uses motion
and grasp planning, reinforcement learning, and self-supervision. We show
significant performance gains over baselines on a simulation benchmark,
sim-to-sim transfer and sim-to-real transfer. |
---|---|
DOI: | 10.48550/arxiv.2303.17592 |