StarVQA: Space-Time Attention for Video Quality Assessment

08/22/2021
by   Fengchuang Xing, et al.
0

The attention mechanism is blooming in computer vision nowadays. However, its application to video quality assessment (VQA) has not been reported. Evaluating the quality of in-the-wild videos is challenging due to the unknown of pristine reference and shooting distortion. This paper presents a novel space-time attention network for the VQA problem, named StarVQA. StarVQA builds a Transformer by alternately concatenating the divided space-time attention. To adapt the Transformer architecture for training, StarVQA designs a vectorized regression loss by encoding the mean opinion score (MOS) to the probability vector and embedding a special vectorized label token as the learnable variable. To capture the long-range spatiotemporal dependencies of a video sequence, StarVQA encodes the space-time position information of each patch to the input of the Transformer. Various experiments are conducted on the de-facto in-the-wild video datasets, including LIVE-VQC, KoNViD-1k, LSVQ, and LSVQ-1080p. Experimental results demonstrate the superiority of the proposed StarVQA over the state-of-the-art. Code and model will be available at: https://github.com/DVL/StarVQA.

READ FULL TEXT

page 2

page 4

research
06/21/2023

StarVQA+: Co-training Space-Time Attention for Video Quality Assessment

Self-attention based Transformer has achieved great success in many comp...
research
07/24/2022

Object State Change Classification in Egocentric Videos using the Divided Space-Time Attention Mechanism

This report describes our submission called "TarHeels" for the Ego4D: Ob...
research
11/27/2020

Patch-VQ: 'Patching Up' the Video Quality Problem

No-reference (NR) perceptual video quality assessment (VQA) is a complex...
research
03/13/2023

MRET: Multi-resolution Transformer for Video Quality Assessment

No-reference video quality assessment (NR-VQA) for user generated conten...
research
09/19/2022

Panoramic Vision Transformer for Saliency Detection in 360° Videos

360^∘ video saliency detection is one of the challenging benchmarks for ...
research
09/17/2021

ChipQA: No-Reference Video Quality Prediction via Space-Time Chips

We propose a new model for no-reference video quality assessment (VQA). ...
research
05/22/2023

Towards Explainable In-the-Wild Video Quality Assessment: A Database and a Language-Prompted Approach

The proliferation of in-the-wild videos has greatly expanded the Video Q...

Please sign up or login with your details

Forgot password? Click here to reset