Visual-Tactile Cross-Modal Data Generation using Residue-Fusion GAN with Feature-Matching and Perceptual Losses

07/12/2021
by   Shaoyu Cai, et al.
5

Existing psychophysical studies have revealed that the cross-modal visual-tactile perception is common for humans performing daily activities. However, it is still challenging to build the algorithmic mapping from one modality space to another, namely the cross-modal visual-tactile data translation/generation, which could be potentially important for robotic operation. In this paper, we propose a deep-learning-based approach for cross-modal visual-tactile data generation by leveraging the framework of the generative adversarial networks (GANs). Our approach takes the visual image of a material surface as the visual data, and the accelerometer signal induced by the pen-sliding movement on the surface as the tactile data. We adopt the conditional-GAN (cGAN) structure together with the residue-fusion (RF) module, and train the model with the additional feature-matching (FM) and perceptual losses to achieve the cross-modal data generation. The experimental results show that the inclusion of the RF module, and the FM and the perceptual losses significantly improves cross-modal data generation performance in terms of the classification accuracy upon the generated data and the visual similarity between the ground-truth and the generated data.

READ FULL TEXT

page 1

page 3

page 5

page 6

page 7

research
02/17/2019

"Touching to See" and "Seeing to Feel": Robotic Cross-modal SensoryData Generation for Visual-Tactile Perception

The integration of visual-tactile stimulus is common while humans perfor...
research
06/14/2019

Connecting Touch and Vision via Cross-Modal Prediction

Humans perceive the world using multi-modal sensory inputs such as visio...
research
01/18/2020

A Transfer Learning Approach to Cross-Modal Object Recognition: From Visual Observation to Robotic Haptic Exploration

In this work, we introduce the problem of cross-modal visuo-tactile obje...
research
04/02/2018

SyncGAN: Synchronize the Latent Space of Cross-modal Generative Adversarial Networks

Generative adversarial network (GAN) has achieved impressive success on ...
research
10/29/2017

A Novel Approach to Artistic Textual Visualization via GAN

While the visualization of statistical data tends to a mature technology...
research
12/28/2021

Multimodal perception for dexterous manipulation

Humans usually perceive the world in a multimodal way that vision, touch...
research
09/27/2021

Audio-to-Image Cross-Modal Generation

Cross-modal representation learning allows to integrate information from...

Please sign up or login with your details

Forgot password? Click here to reset