Text2Action: Generative Adversarial Synthesis from Language to Action

10/15/2017
by   Hyemin Ahn, et al.
0

In this paper, we propose a generative model which learns the relationship between language and human action in order to generate a human action sequence given a sentence describing human behavior. The proposed generative model is a generative adversarial network (GAN), which is based on the sequence to sequence (SEQ2SEQ) model. Using the proposed generative network, we can synthesize various actions for a robot or a virtual agent using a text encoder recurrent neural network (RNN) and an action decoder RNN. The proposed generative network is trained from 29,770 pairs of actions and sentence annotations extracted from MSR-Video-to-Text (MSR-VTT), a large-scale video dataset. We demonstrate that the network can generate human-like actions which can be transferred to a Baxter robot, such that the robot performs an action based on a provided sentence. Results show that the proposed generative network correctly models the relationship between language and action and can generate a diverse set of actions from the same sentence.

READ FULL TEXT

page 4

page 5

page 7

page 8

research
05/26/2018

Human Action Generation with Generative Adversarial Networks

Inspired by the recent advances in generative models, we introduce a hum...
research
10/15/2021

Pose-guided Generative Adversarial Net for Novel View Action Synthesis

We focus on the problem of novel-view human action synthesis. Given an a...
research
03/24/2023

Factor Decomposed Generative Adversarial Networks for Text-to-Image Synthesis

Prior works about text-to-image synthesis typically concatenated the sen...
research
03/23/2020

Caption Generation of Robot Behaviors based on Unsupervised Learning of Action Segments

Bridging robot action sequences and their natural language captions is a...
research
02/14/2019

Actions Generation from Captions

Sequence transduction models have been widely explored in many natural l...
research
04/13/2022

Controllable Video Generation through Global and Local Motion Dynamics

We present GLASS, a method for Global and Local Action-driven Sequence S...
research
06/24/2018

Generative Models for Pose Transfer

We investigate nearest neighbor and generative models for transferring p...

Please sign up or login with your details

Forgot password? Click here to reset