Graph-based Spatial-temporal Feature Learning for Neuromorphic Vision Sensing

10/08/2019
by   Yin Bi, et al.
0

Neuromorphic vision sensing (NVS) allows for significantly higher event sampling rates at substantially increased energy efficiency and robustness to illumination changes. However, feature representation for NVS is far behind the APS-based counterparts, resulting in lower performance in high-level computer vision tasks. To fully utilize the sparse and asynchronous nature, we propose a compact graph representation for NVS, which allows for end-to-end learning with graph convolution neural networks. We couple this with a novel end-to-end feature learning framework that accommodates both appearance-based and motion-based tasks. The core of framework comprises a spatial feature learning module, which utilizes our proposed residual-graph CNN (RG-CNN), for end-to-end learning of appearance-based features directly from graphs. We extend this with our proposed Graph2Grid block and temporal feature learning module for efficiently modelling temporal dependencies over multiple graphs and a long temporal extent. We show that performance of this framework generalizes to both object classification and action recognition, which, importantly, preserves the spatial and temporal coherence of spike events, while requiring less computation and memory. The experimental validation shows that our framework show that our proposed framework outperforms all recent methods on standard datasets. Finally, to address the absence of large real-world NVS datasets for complex recognition tasks, we introduce, evaluate and make available a 100k dataset of NVS recordings of the American Sign Language letters (ASL_DVS) acquired with an iniLabs DAVIS240c device under real-world conditions, as well as a neuromorphic action recognition dataset (UCF101_DVS and HMDB51_DVS) recorded from monitor.

READ FULL TEXT

page 4

page 11

research
08/19/2019

Graph-Based Object Classification for Neuromorphic Vision Sensing

Neuromorphic vision sensing (NVS) devices represent visual information a...
research
03/04/2019

Collaborative Spatio-temporal Feature Learning for Video Action Recognition

Spatio-temporal feature learning is of central importance for action rec...
research
07/12/2017

Capacity, Fidelity, and Noise Tolerance of Associative Spatial-Temporal Memories Based on Memristive Neuromorphic Network

We have calculated the key characteristics of associative (content-addre...
research
09/28/2020

Event-based Action Recognition Using Timestamp Image Encoding Network

Event camera is an asynchronous, high frequency vision sensor with low p...
research
09/02/2020

IAUnet: Global Context-Aware Feature Learning for Person Re-Identification

Person re-identification (reID) by CNNs based networks has achieved favo...
research
09/23/2020

Unsupervised Feature Learning for Event Data: Direct vs Inverse Problem Formulation

Event-based cameras record an asynchronous stream of per-pixel brightnes...
research
11/24/2017

Appearance-and-Relation Networks for Video Classification

Spatiotemporal feature learning in videos is a fundamental and difficult...

Please sign up or login with your details

Forgot password? Click here to reset