Qiniu Submission to ActivityNet Challenge 2018

06/12/2018
by   Xiaoteng Zhang, et al.
0

In this paper, we introduce our submissions for the tasks of trimmed activity recognition (Kinetics) and trimmed event recognition (Moments in Time) for Activitynet Challenge 2018. In the two tasks, non-local neural networks and temporal segment networks are implemented as our base models. Multi-modal cues such as RGB image, optical flow and acoustic signal have also been used in our method. We also propose new non-local-based models for further improvement on the recognition accuracy. The final submissions after ensembling the models achieve 83.5 validation set, 35.81 validation set.

READ FULL TEXT
research
08/03/2021

Non-local Graph Convolutional Network for joint Activity Recognition and Motion Prediction

3D skeleton-based motion prediction and activity recognition are two int...
research
07/02/2018

Multi-modal Egocentric Activity Recognition using Audio-Visual Features

Egocentric activity recognition in first-person videos has an increasing...
research
03/12/2020

Top-1 Solution of Multi-Moments in Time Challenge 2019

In this technical report, we briefly introduce the solutions of our team...
research
06/27/2018

Exploiting Spatial-Temporal Modelling and Multi-Modal Fusion for Human Action Recognition

In this report, our approach to tackling the task of ActivityNet 2018 Ki...
research
11/07/2020

Non-local convolutional neural networks (nlcnn) for speaker recognition

Speaker recognition is the process of identifying a speaker based on the...
research
12/04/2019

Template co-updating in multi-modal human activity recognition systems

Multi-modal systems are quite common in the context of human activity re...
research
10/31/2018

GraphIE: A Graph-Based Framework for Information Extraction

Most modern Information Extraction (IE) systems are implemented as seque...

Please sign up or login with your details

Forgot password? Click here to reset