GI-NNet & RGI-NNet: Development of Robotic Grasp Pose Models, Trainable with Large as well as Limited Labelled Training Datasets, under supervised and semi supervised paradigms

07/15/2021
by   Priya Shukla, et al.
4

Our way of grasping objects is challenging for efficient, intelligent and optimal grasp by COBOTs. To streamline the process, here we use deep learning techniques to help robots learn to generate and execute appropriate grasps quickly. We developed a Generative Inception Neural Network (GI-NNet) model, capable of generating antipodal robotic grasps on seen as well as unseen objects. It is trained on Cornell Grasping Dataset (CGD) and attained 98.87 grasp pose accuracy for detecting both regular and irregular shaped objects from RGB-Depth (RGB-D) images while requiring only one third of the network trainable parameters as compared to the existing approaches. However, to attain this level of performance the model requires the entire 90 labelled data of CGD keeping only 10 vulnerable to poor generalization. Furthermore, getting sufficient and quality labelled dataset is becoming increasingly difficult keeping in pace with the requirement of gigantic networks. To address these issues, we attach our model as a decoder with a semi-supervised learning based architecture known as Vector Quantized Variational Auto Encoder (VQVAE), which works efficiently when trained both with the available labelled and unlabelled data. The proposed model, which we name as Representation based GI-NNet (RGI-NNet), has been trained with various splits of label data on CGD with as minimum as 10 labelled dataset together with latent embedding generated from VQVAE up to 50 labelled data with latent embedding obtained from VQVAE. The performance level, in terms of grasp pose accuracy of RGI-NNet, varies between 92.13 which is far better than several existing models trained with only labelled dataset. For the performance verification of both GI-NNet and RGI-NNet models, we use Anukul (Baxter) hardware cobot.

READ FULL TEXT

page 1

page 7

page 8

page 9

research
02/20/2022

Generating Quality Grasp Rectangle using Pix2Pix GAN for Intelligent Robot Grasping

Intelligent robot grasping is a very challenging task due to its inheren...
research
01/23/2020

Semi-supervised Grasp Detection by Representation Learning in a Vector Quantized Latent Space

Determining quality grasps from an image is an important area of researc...
research
08/07/2016

Deep Learning a Grasp Function for Grasping under Gripper Pose Uncertainty

This paper presents a new method for parallel-jaw grasping of isolated o...
research
04/04/2018

Generative Visual Rationales

Interpretability and small labelled datasets are key issues in the pract...
research
10/21/2022

Auto-Encoder Neural Network Incorporating X-Ray Fluorescence Fundamental Parameters with Machine Learning

We consider energy-dispersive X-ray Fluorescence (EDXRF) applications wh...
research
08/22/2023

Vision-Based Intelligent Robot Grasping Using Sparse Neural Network

In the modern era of Deep Learning, network parameters play a vital role...
research
12/10/2021

The Large Labelled Logo Dataset (L3D): A Multipurpose and Hand-Labelled Continuously Growing Dataset

In this work, we present the Large Labelled Logo Dataset (L3D), a multip...

Please sign up or login with your details

Forgot password? Click here to reset