Combining Events and Frames using Recurrent Asynchronous Multimodal Networks for Monocular Depth Prediction

02/18/2021
by   Daniel Gehrig, et al.
9

Event cameras are novel vision sensors that report per-pixel brightness changes as a stream of asynchronous "events". They offer significant advantages compared to standard cameras due to their high temporal resolution, high dynamic range and lack of motion blur. However, events only measure the varying component of the visual signal, which limits their ability to encode scene context. By contrast, standard cameras measure absolute intensity frames, which capture a much richer representation of the scene. Both sensors are thus complementary. However, due to the asynchronous nature of events, combining them with synchronous images remains challenging, especially for learning-based methods. This is because traditional recurrent neural networks (RNNs) are not designed for asynchronous and irregular data from additional sensors. To address this challenge, we introduce Recurrent Asynchronous Multimodal (RAM) networks, which generalize traditional RNNs to handle asynchronous and irregular data from multiple sensors. Inspired by traditional RNNs, RAM networks maintain a hidden state that is updated asynchronously and can be queried at any time to generate a prediction. We apply this novel architecture to monocular depth estimation with events and frames where we show an improvement over state-of-the-art methods by up to 30 absolute depth error. To enable further research on multimodal learning with events, we release EventScape, a new dataset with events, intensity frames, semantic labels, and depth maps recorded in the CARLA simulator.

READ FULL TEXT

page 1

page 5

page 7

page 9

page 10

research
10/16/2020

Learning Monocular Dense Depth from Events

Event cameras are novel sensors that output brightness changes in the fo...
research
04/17/2019

Events-to-Video: Bringing Modern Computer Vision to Event Cameras

Event cameras are novel sensors that report brightness changes in the fo...
research
07/25/2018

Asynchronous, Photometric Feature Tracking using Events and Frames

We present a method that leverages the complementarity of event cameras ...
research
10/17/2022

Event-based Stereo Depth Estimation from Ego-motion using Ray Density Fusion

Event cameras are bio-inspired sensors that mimic the human retina by re...
research
12/10/2020

An Asynchronous Kalman Filter for Hybrid Event Cameras

We present an Asynchronous Kalman Filter (AKF) to reconstruct High Dynam...
research
09/03/2023

An Asynchronous Linear Filter Architecture for Hybrid Event-Frame Cameras

Event cameras are ideally suited to capture High Dynamic Range (HDR) vis...
research
09/04/2021

Dual Transfer Learning for Event-based End-task Prediction via Pluggable Event to Image Translation

Event cameras are novel sensors that perceive the per-pixel intensity ch...

Please sign up or login with your details

Forgot password? Click here to reset