Visuo-Tactile Transformers for Manipulation

09/30/2022
by   Yizhou Chen, et al.
0

Learning representations in the joint domain of vision and touch can improve manipulation dexterity, robustness, and sample-complexity by exploiting mutual information and complementary cues. Here, we present Visuo-Tactile Transformers (VTTs), a novel multimodal representation learning approach suited for model-based reinforcement learning and planning. Our approach extends the Visual Transformer <cit.> to handle visuo-tactile feedback. Specifically, VTT uses tactile feedback together with self and cross-modal attention to build latent heatmap representations that focus attention on important task features in the visual domain. We demonstrate the efficacy of VTT for representation learning with a comparative evaluation against baselines on four simulated robot tasks and one real world block pushing task. We conduct an ablation study over the components of VTT to highlight the importance of cross-modality in representation learning.

READ FULL TEXT

page 6

page 8

page 13

page 14

research
09/26/2022

Learning Self-Supervised Representations from Vision and Touch for Active Sliding Perception of Deformable Surfaces

Humans make extensive use of vision and touch as complementary senses, w...
research
12/28/2021

Multimodal perception for dexterous manipulation

Humans usually perceive the world in a multimodal way that vision, touch...
research
06/23/2016

Unsupervised preprocessing for Tactile Data

Tactile information is important for gripping, stable grasp, and in-hand...
research
10/07/2022

VIRDO++: Real-World, Visuo-tactile Dynamics and Perception of Deformable Objects

Deformable objects manipulation can benefit from representations that se...
research
12/08/2020

Parameter Efficient Multimodal Transformers for Video Representation Learning

The recent success of Transformers in the language domain has motivated ...
research
01/21/2021

Learning rich touch representations through cross-modal self-supervision

The sense of touch is fundamental in several manipulation tasks, but rar...
research
07/14/2022

Visuo-Tactile Manipulation Planning Using Reinforcement Learning with Affordance Representation

Robots are increasingly expected to manipulate objects in ever more unst...

Please sign up or login with your details

Forgot password? Click here to reset