Self-Supervised Learning of Action Affordances as Interaction Modes

05/27/2023
by   Liquan Wang, et al.
0

When humans perform a task with an articulated object, they interact with the object only in a handful of ways, while the space of all possible interactions is nearly endless. This is because humans have prior knowledge about what interactions are likely to be successful, i.e., to open a new door we first try the handle. While learning such priors without supervision is easy for humans, it is notoriously hard for machines. In this work, we tackle unsupervised learning of priors of useful interactions with articulated objects, which we call interaction modes. In contrast to the prior art, we use no supervision or privileged information; we only assume access to the depth sensor in the simulator to learn the interaction modes. More precisely, we define a successful interaction as the one changing the visual environment substantially and learn a generative model of such interactions, that can be conditioned on the desired goal state of the object. In our experiments, we show that our model covers most of the human interaction modes, outperforms existing state-of-the-art methods for affordance learning, and can generalize to objects never seen during training. Additionally, we show promising results in the goal-conditional setup, where our model can be quickly fine-tuned to perform a given task. We show in the experiments that such affordance learning predicts interaction which covers most modes of interaction for the querying articulated object and can be fine-tuned to a goal-conditional model. For supplementary: https://actaim.github.io.

READ FULL TEXT

page 1

page 3

page 4

page 5

page 6

research
09/06/2022

Reconstructing Action-Conditioned Human-Object Interactions Using Commonsense Knowledge Priors

We present a method for inferring diverse 3D models of human-object inte...
research
08/23/2023

CHORUS: Learning Canonicalized 3D Human-Object Spatial Relations from Unbounded Synthesized Images

We present a method for teaching machines to understand and model the un...
research
09/27/2020

Human-Object Interaction Detection:A Quick Survey and Examination of Methods

Human-object interaction detection is a relatively new task in the world...
research
07/26/2022

Compositional Human-Scene Interaction Synthesis with Semantic Control

Synthesizing natural interactions between virtual humans and their 3D en...
research
12/01/2021

Human-Object Interaction Detection via Weak Supervision

The goal of this paper is Human-object Interaction (HO-I) detection. HO-...
research
06/16/2020

Learning About Objects by Learning to Interact with Them

Much of the remarkable progress in computer vision has been focused arou...
research
03/03/2020

Self-Supervised Object-Level Deep Reinforcement Learning

Current deep reinforcement learning approaches incorporate minimal prior...

Please sign up or login with your details

Forgot password? Click here to reset