On the Choice of Perception Loss Function for Learned Video Compression

05/30/2023
by   Sadaf Salehkalaibar, et al.
0

We study causal, low-latency, sequential video compression when the output is subjected to both a mean squared-error (MSE) distortion loss as well as a perception loss to target realism. Motivated by prior approaches, we consider two different perception loss functions (PLFs). The first, PLF-JD, considers the joint distribution (JD) of all the video frames up to the current one, while the second metric, PLF-FMD, considers the framewise marginal distributions (FMD) between the source and reconstruction. Using information theoretic analysis and deep-learning based experiments, we demonstrate that the choice of PLF can have a significant effect on the reconstruction, especially at low-bit rates. In particular, while the reconstruction based on PLF-JD can better preserve the temporal correlation across frames, it also imposes a significant penalty in distortion compared to PLF-FMD and further makes it more difficult to recover from errors made in the earlier output frames. Although the choice of PLF decisively affects reconstruction quality, we also demonstrate that it may not be essential to commit to a particular PLF during encoding and the choice of PLF can be delegated to the decoder. In particular, encoded representations generated by training a system to minimize the MSE (without requiring either PLF) can be near universal and can generate close to optimal reconstructions for either choice of PLF at the decoder. We validate our results using (one-shot) information-theoretic analysis, detailed study of the rate-distortion-perception tradeoff of the Gauss-Markov source model as well as deep-learning based experiments on moving MNIST and KTH datasets.

READ FULL TEXT

page 2

page 8

research
06/18/2021

Universal Rate-Distortion-Perception Representations for Lossy Compression

In the context of lossy compression, Blau Michaeli (2019) adopt a ma...
research
12/28/2022

Multi-Realism Image Compression with a Conditional Generator

By optimizing the rate-distortion-realism trade-off, generative compress...
research
05/24/2019

A Compression Objective and a Cycle Loss for Neural Image Compression

In this manuscript we propose two objective terms for neural image compr...
research
08/12/2022

Semantic Compression with Side Information: A Rate-Distortion Perspective

We consider the semantic rate-distortion problem motivated by task-orien...
research
02/14/2023

Cauchy Loss Function: Robustness Under Gaussian and Cauchy Noise

In supervised machine learning, the choice of loss function implicitly a...
research
09/06/2023

EGIC: Enhanced Low-Bit-Rate Generative Image Compression Guided by Semantic Segmentation

We introduce EGIC, a novel generative image compression method that allo...
research
09/21/2020

Integration of Clinical Criteria into the Training of Deep Models: Application to Glucose Prediction for Diabetic People

Standard objective functions used during the training of neural-network-...

Please sign up or login with your details

Forgot password? Click here to reset