When Transformer Meets Robotic Grasping: Exploits Context for Efficient Grasp Detection

02/24/2022
by   Shaochen Wang, et al.
0

In this paper, we present a transformer-based architecture, namely TF-Grasp, for robotic grasp detection. The developed TF-Grasp framework has two elaborate designs making it well suitable for visual grasping tasks. The first key design is that we adopt the local window attention to capture local contextual information and detailed features of graspable objects. Then, we apply the cross window attention to model the long-term dependencies between distant pixels. Object knowledge, environmental configuration, and relationships between different visual entities are aggregated for subsequent grasp detection. The second key design is that we build a hierarchical encoder-decoder architecture with skip-connections, delivering shallow features from encoder to decoder to enable a multi-scale feature fusion. Due to the powerful attention mechanism, the TF-Grasp can simultaneously obtain the local information (i.e., the contours of objects), and model long-term connections such as the relationships between distinct visual concepts in clutter. Extensive computational experiments demonstrate that the TF-Grasp achieves superior results versus state-of-art grasping convolutional models and attain a higher accuracy of 97.99 respectively. Real-world experiments using a 7DoF Franka Emika Panda robot also demonstrate its capability of grasping unseen objects in a variety of scenarios. The code and pre-trained models will be available at https://github.com/WangShaoSUN/grasp-transformer

READ FULL TEXT

page 1

page 3

page 5

page 6

page 7

research
05/30/2022

Robotic grasp detection based on Transformer

Grasp detection in a cluttered environment is still a great challenge fo...
research
01/29/2023

6-DoF Robotic Grasping with Transformer

Robotic grasping aims to detect graspable points and their corresponding...
research
09/13/2022

What You See is What You Grasp: User-Friendly Grasping Guided by Near-eye-tracking

This work presents a next-generation human-robot interface that can infe...
research
06/17/2023

NBMOD: Find It and Grasp It in Noisy Background

Grasping objects is a fundamental yet important capability of robots, an...
research
09/08/2022

A Secure and Efficient Multi-Object Grasping Detection Approach for Robotic Arms

Robotic arms are widely used in automatic industries. However, with wide...
research
12/10/2022

Towards Scale Balanced 6-DoF Grasp Detection in Cluttered Scenes

In this paper, we focus on the problem of feature learning in the presen...
research
08/02/2023

Grasp Stability Assessment Through Attention-Guided Cross-Modality Fusion and Transfer Learning

Extensive research has been conducted on assessing grasp stability, a cr...

Please sign up or login with your details

Forgot password? Click here to reset