Perceive, Represent, Generate: Translating Multimodal Information to Robotic Motion Trajectories

04/06/2022
by   Fábio Vital, et al.
0

We present Perceive-Represent-Generate (PRG), a novel three-stage framework that maps perceptual information of different modalities (e.g., visual or sound), corresponding to a sequence of instructions, to an adequate sequence of movements to be executed by a robot. In the first stage, we perceive and pre-process the given inputs, isolating individual commands from the complete instruction provided by a human user. In the second stage we encode the individual commands into a multimodal latent space, employing a deep generative model. Finally, in the third stage we convert the multimodal latent values into individual trajectories and combine them into a single dynamic movement primitive, allowing its execution in a robotic platform. We evaluate our pipeline in the context of a novel robotic handwriting task, where the robot receives as input a word through different perceptual modalities (e.g., image, sound), and generates the corresponding motion trajectory to write it, creating coherent and readable handwritten words.

READ FULL TEXT

page 4

page 6

research
04/20/2022

Sound-Guided Semantic Video Generation

The recent success in StyleGAN demonstrates that pre-trained StyleGAN la...
research
05/25/2023

Score-Based Multimodal Autoencoders

Multimodal Variational Autoencoders (VAEs) represent a promising group o...
research
02/09/2023

Robot Synesthesia: A Sound and Emotion Guided AI Painter

If a picture paints a thousand words, sound may voice a million. While r...
research
10/09/2019

Multimodal representation models for prediction and control from partial information

Similar to humans, robots benefit from interacting with their environmen...
research
11/09/2020

Reward Conditioned Neural Movement Primitives for Population Based Variational Policy Optimization

The aim of this paper is to study the reward based policy exploration pr...
research
05/18/2018

Learning and Inference Movement with Deep Generative Model

Learning and inference movement is a very challenging problem due to its...
research
10/15/2020

A Reversible Dynamic Movement Primitive formulation

In this work, a novel Dynamic Movement Primitive (DMP) formulation is pr...

Please sign up or login with your details

Forgot password? Click here to reset