A Deep Moving-camera Background Model

09/16/2022
by   Guy Erez, et al.
25

In video analysis, background models have many applications such as background/foreground separation, change detection, anomaly detection, tracking, and more. However, while learning such a model in a video captured by a static camera is a fairly-solved task, in the case of a Moving-camera Background Model (MCBM), the success has been far more modest due to algorithmic and scalability challenges that arise due to the camera motion. Thus, existing MCBMs are limited in their scope and their supported camera-motion types. These hurdles also impeded the employment, in this unsupervised task, of end-to-end solutions based on deep learning (DL). Moreover, existing MCBMs usually model the background either on the domain of a typically-large panoramic image or in an online fashion. Unfortunately, the former creates several problems, including poor scalability, while the latter prevents the recognition and leveraging of cases where the camera revisits previously-seen parts of the scene. This paper proposes a new method, called DeepMCBM, that eliminates all the aforementioned issues and achieves state-of-the-art results. Concretely, first we identify the difficulties associated with joint alignment of video frames in general and in a DL setting in particular. Next, we propose a new strategy for joint alignment that lets us use a spatial transformer net with neither a regularization nor any form of specialized (and non-differentiable) initialization. Coupled with an autoencoder conditioned on unwarped robust central moments (obtained from the joint alignment), this yields an end-to-end regularization-free MCBM that supports a broad range of camera motions and scales gracefully. We demonstrate DeepMCBM's utility on a variety of videos, including ones beyond the scope of other methods. Our code is available at https://github.com/BGU-CS-VIL/DeepMCBM .

READ FULL TEXT

page 2

page 6

page 9

page 10

page 13

research
02/10/2022

Dynamic Background Subtraction by Generative Neural Networks

Background subtraction is a significant task in computer vision and an e...
research
03/06/2019

Robust Video Background Identification by Dominant Rigid Motion Estimation

The ability to identify the static background in videos captured by a mo...
research
06/06/2023

YONA: You Only Need One Adjacent Reference-frame for Accurate and Fast Video Polyp Detection

Accurate polyp detection is essential for assisting clinical rectal canc...
research
01/07/2018

Foreground Segmentation Using a Triplet Convolutional Neural Network for Multiscale Feature Encoding

A common approach for moving objects segmentation in a scene is to perfo...
research
03/18/2016

Approximated Robust Principal Component Analysis for Improved General Scene Background Subtraction

The research reported in this paper addresses the fundamental task of se...
research
03/30/2022

End to End Lip Synchronization with a Temporal AutoEncoder

We study the problem of syncing the lip movement in a video with the aud...
research
10/15/2009

Behavior Subtraction

Background subtraction has been a driving engine for many computer visio...

Please sign up or login with your details

Forgot password? Click here to reset