SLIC: Self-Conditioned Adaptive Transform with Large-Scale Receptive Fields for Learned Image Compression

04/19/2023
by   Wei Jiang, et al.
0

Learned image compression has achieved remarkable performance. Transform, plays an important role in boosting the RD performance. Analysis transform converts the input image to a compact latent representation. The more compact the latent representation is, the fewer bits we need to compress it. When designing better transform, some previous works adopt Swin-Transformer. The success of the Swin-Transformer in image compression can be attributed to the dynamic weights and large receptive field.However,the LayerNorm adopted in transformers is not suitable for image compression.We find CNN-based modules can also be dynamic and have large receptive-fields. The CNN-based modules can also work with GDN/IGDN. To make the CNN-based modules dynamic, we generate the weights of kernels conditioned on the input feature. We scale up the size of each kernel for larger receptive fields. To reduce complexity, we make the CNN-module channel-wise connected. We call this module Dynamic Depth-wise convolution. We replace the self-attention module with the proposed Dynamic Depth-wise convolution, replace the embedding layer with a depth-wise residual bottleneck for non-linearity and replace the FFN layer with an inverted residual bottleneck for more interactions in the spatial domain. The interactions among channels of dynamic depth-wise convolution are limited. We design the other block, which replaces the dynamic depth-wise convolution with channel attention. We equip the proposed modules in the analysis and synthesis transform and receive a more compact latent representation and propose the learned image compression model SLIC, meaning Self-Conditioned Adaptive Transform with Large-Scale Receptive Fields for Learned Image Compression Learned Image Compression. Thanks to the proposed transform modules, our proposed SLIC achieves 6.35 on Kodak dataset.

READ FULL TEXT

page 1

page 4

page 9

research
08/17/2023

Dynamic Kernel-Based Adaptive Spatial Aggregation for Learned Image Compression

Learned image compression methods have shown superior rate-distortion pe...
research
03/27/2023

Learned Image Compression with Mixed Transformer-CNN Architectures

Learned image compression (LIC) methods have exhibited promising progres...
research
08/14/2023

SCSC: Spatial Cross-scale Convolution Module to Strengthen both CNNs and Transformers

This paper presents a module, Spatial Cross-scale Convolution (SCSC), wh...
research
09/04/2023

Large Separable Kernel Attention: Rethinking the Large Kernel Attention Design in CNN

Visual Attention Networks (VAN) with Large Kernel Attention (LKA) module...
research
08/10/2022

Multi-scale Feature Aggregation for Crowd Counting

Convolutional Neural Network (CNN) based crowd counting methods have ach...
research
08/05/2022

Blockwise Temporal-Spatial Pathway Network

Algorithms for video action recognition should consider not only spatial...
research
04/25/2022

High-Efficiency Lossy Image Coding Through Adaptive Neighborhood Information Aggregation

Questing for lossy image coding (LIC) with superior efficiency on both c...

Please sign up or login with your details

Forgot password? Click here to reset