Hallucinating Optical Flow Features for Video Classification

05/28/2019
by   Yongyi Tang, et al.
0

Appearance and motion are two key components to depict and characterize the video content. Currently, the two-stream models have achieved state-of-the-art performances on video classification. However, extracting motion information, specifically in the form of optical flow features, is extremely computationally expensive, especially for large-scale video classification. In this paper, we propose a motion hallucination network, namely MoNet, to imagine the optical flow features from the appearance features, with no reliance on the optical flow computation. Specifically, MoNet models the temporal relationships of the appearance features and exploits the contextual relationships of the optical flow features with concurrent connections. Extensive experimental results demonstrate that the proposed MoNet can effectively and efficiently hallucinate the optical flow features, which together with the appearance features consistently improve the video classification performances. Moreover, MoNet can help cutting down almost a half of computational and data-storage burdens for the two-stream video classification. Our code is available at: https://github.com/YongyiTang92/MoNet-Features.

READ FULL TEXT
research
11/20/2021

FAMINet: Learning Real-time Semi-supervised Video Object Segmentation with Steepest Optimized Optical Flow

Semi-supervised video object segmentation (VOS) aims to segment a few mo...
research
04/29/2021

AutoFlow: Learning a Better Training Set for Optical Flow

Synthetic datasets play a critical role in pre-training CNN models for o...
research
09/15/2023

Privacy-preserving Early Detection of Epileptic Seizures in Videos

In this work, we contribute towards the development of video-based epile...
research
08/31/2016

Efficient Two-Stream Motion and Appearance 3D CNNs for Video Classification

The video and action classification have extremely evolved by deep neura...
research
12/30/2017

A Unified Method for First and Third Person Action Recognition

In this paper, a new video classification methodology is proposed which ...
research
11/21/2019

MIMAMO Net: Integrating Micro- and Macro-motion for Video Emotion Recognition

Spatial-temporal feature learning is of vital importance for video emoti...
research
02/02/2015

Learning the Matching Function

The matching function for the problem of stereo reconstruction or optica...

Please sign up or login with your details

Forgot password? Click here to reset