VSGM – Enhance robot task understanding ability through visual semantic graph

05/19/2021
by   Cheng-Yu Tsai, et al.
15

In recent years, developing AI for robotics has raised much attention. The interaction of vision and language of robots is particularly difficult. We consider that giving robots an understanding of visual semantics and language semantics will improve inference ability. In this paper, we propose a novel method-VSGM (Visual Semantic Graph Memory), which uses the semantic graph to obtain better visual image features, improve the robot's visual understanding ability. By providing prior knowledge of the robot and detecting the objects in the image, it predicts the correlation between the attributes of the object and the objects and converts them into a graph-based representation; and mapping the object in the image to be a top-down egocentric map. Finally, the important object features of the current task are extracted by Graph Neural Networks. The method proposed in this paper is verified in the ALFRED (Action Learning From Realistic Environments and Directives) dataset. In this dataset, the robot needs to perform daily indoor household tasks following the required language instructions. After the model is added to the VSGM, the task success rate can be improved by 6 10

READ FULL TEXT

page 3

page 4

page 9

page 15

page 16

research
01/11/2023

Graph based Environment Representation for Vision-and-Language Navigation in Continuous Environments

Vision-and-Language Navigation in Continuous Environments (VLN-CE) is a ...
research
07/25/2021

Improving Robot Localisation by Ignoring Visual Distraction

Attention is an important component of modern deep learning. However, le...
research
10/16/2017

Pushing the envelope in deep visual recognition for mobile platforms

Image classification is the task of assigning to an input image a label ...
research
06/09/2016

Understanding User Instructions by Utilizing Open Knowledge for Service Robots

Understanding user instructions in natural language is an active researc...
research
09/16/2019

Constructing Dynamic Knowledge Graph for Visual Semantic Understanding and Applications in Autonomous Robotics

Interpreting semantic knowledge describing entities, relations and attri...
research
04/09/2022

Unbiased Directed Object Attention Graph for Object Navigation

Object navigation tasks require agents to locate specific objects in unk...
research
04/18/2018

Object Ordering with Bidirectional Matchings for Visual Reasoning

Visual reasoning with compositional natural language instructions, e.g.,...

Please sign up or login with your details

Forgot password? Click here to reset