An Autoencoder-based Learned Image Compressor: Description of Challenge Proposal by NCTU

02/20/2019 ∙ by David Alexandre, et al. ∙ National Chiao Tung University 0

We propose a lossy image compression system using the deep-learning autoencoder structure to participate in the Challenge on Learned Image Compression (CLIC) 2018. Our autoencoder uses the residual blocks with skip connections to reduce the correlation among image pixels and condense the input image into a set of feature maps, a compact representation of the original image. The bit allocation and bitrate control are implemented by using the importance maps and quantizer. The importance maps are generated by a separate neural net in the encoder. The autoencoder and the importance net are trained jointly based on minimizing a weighted sum of mean squared error, MS-SSIM, and a rate estimate. Our aim is to produce reconstructed images with good subjective quality subject to the 0.15 bits-per-pixel constraint.



There are no comments yet.


page 2

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Lossy image compression has been a challenging topic in deep learning domain in recent years. Until now, several methods have been proposed for compressing images using deep learning techniques, for example, end-to-end trained neural networks, post-processing neural networks with non-neural codecs, or improving traditional codecs using neural networks. The autoencoder structure has been a popular choice to do end-to-end image compression. The encoder part learns to generate a compact representation of the input image, and the decoder part reconstructs an image close to the input based on the compact representation. Similar to the traditional techniques, the performance of a lossy image compression scheme relies on how well the quantization and rate control scheme are blended into this autoencoder structure.

Our response to this Challenge on Learned Image Compression (CLIC) at CVPR 2018 is an end-to-end trainable autoencoder with residual blocks and importance maps. More specifically, our system aims to meet the 0.15 bits-per-pixel requirement while achieving good Peak Signal Noise Ratio (PSNR) performance and subjective quality. Our neural net architecture is inspired by the works of [1] and [2]. We adopt the autoencoder architecture from Mentzer et al. [1] and the importance map concept from Li et al. [2]. The architecture from Mentzer et al. consists of a set of symmetrical convolution layers and residual block layers to compress and decompress an image. The encoder reduces the dimensionality of an image into a number of feature maps. The feature maps contain densely packed color values from an image and can be reconstructed closely into its original value by using the associated decoder. Each feature map is further modulated with an importance map, which controls the bit allocation among feature samples in a feature map, determining critically the final compression rate of the output image. The importance maps are generated by a set of trained convolution layers that can identify the important feature samples (i.e., samples that need more bits) of an image and produce a set of importance masks to truncate the (bit) depth of feature maps in a spatially varying manner. Therefore, the final feature maps keep only the most meaningful information of representing an image at a specific rate. They are essentially the compact representation of the input image.

To overcome the training problem due to the quantization operation, we adopt the soft quantization technique in [3] and implement a straightforward rate estimation based on the values of importance masks for the rate control purpose. Our entropy coder, which compresses the truncated feature maps, is developed based on the JPEG2000 standard. We adjust the context model (in the arithmetic coder) to match our feature map characteristics, which contains all zero at the least significant bitplane in most of the cases.

Figure 1: The proposed network architecture.

As instructed by the challenge guidelines, PSNR is used to measure the output image quality. We compare it with the popular image compression standards such as JPEG and BPG. Also, we use MS-SSIM, which mimics human subjective quality evaluation, as the second image quality metric.

2 Methods

We give more details of our implementation in this section, which includes the deep learning network architecture, the generation of importance map, and finally, the quantization and arithmetic coder in our system.

2.1 Network Architecture

The autoencoder part for feature map generation and image reconstruction is adapted from [1]. The input to the encoder is an image with dimension HW in RGB format. The upper part represents the encoder and the lower part represents the decoder

. k5 n64 s2 represents kernel size 5, with 64 channels and stride 2. All layers in the encoder are the linear convolution and non-linear activation operations, and the decoder layers implement the inverse operations.

Given an input image , the encoder produces a set of feature maps . The normalization at the encoder limits the image input value to . The encoder output , limited in the range of 0 to 1, goes to the quantizer and the importance net. The quantizer, , produces a fixed-point representation of by keeping the first bits after the decimal point. Essentially, this amounts to performing a uniform quantization on with a step size of . In the meanwhile, is the input to the importance net, , and it generates an importance map, , for each feature map. In particular, the values of the importance maps are positive integers ranging from 1 to , indicating how many (most significant) bits of the quantized feature should be kept via the masking operation . The decoder, , further takes in after the binary arithmetic decoding of the bitstream to generate the reconstructed image . The denormalization at the end of decoder reverses the encoder input normalization, clipping the output pixel value to .

2.2 Quantizer

The quantizer is a critical element in lossy image compression. Our quantizer is located at the end of encoder. Specifically, the quantized value of a feature sample is obtained as follows:


where is the desired bit depth (which is 4 in our submitted codec), denotes the floor function, and is the feature sample. In the training phase, we use the following soft-quantization [3] in place of the hard quantization in (1) to facilitate the gradient calculation in back-propagation:


Essentially, this soft-quantization approximates the quantizer mapping function of the hard-quantization with a smooth differentiable function so as to have non-zero gradients at training time. In addition, we save the quantized values with a series of integer indices to save space when storing the multiple quantized values. These indices can be represented in binary form to match the masking operation described in the next subsection.

2.3 Importance Net

Figure 2: Importance maps (left) for an input image (right).

The importance map addresses the issue of content complexity variation in different parts of an image. Also, some images are easier to compress than others. It is closely related to measuring the complexity of an image. Li et al. [2] introduce the importance map as the solution to this problem. It is expected that the importance map, produced by a neural network, can learn to signify the complexity of different areas in an image. Thus, at a given bitrate, the compression system allocates more bits to the complicated areas.

Our importance net is made of residual blocks, followed by a quantizer to produce an integer-valued importance map for each feature map. The acts as a mask to set zero bits of the quantized feature which are less significant in contributing to the overall reconstruction quality, as done by (3) and illustrated in Figure 2. In other words, for each sample on the feature map, the corresponding value, , determines how many bits of information to retain on the final feature map. As such, the importance maps have a resolution and a channel number identical to the feature maps. It is worth noting that they are produced based on the same encoder output . In the training process, the quantizer in the importance net is replaced by the soft-quantization in [3]. Figure 2 shows the importance maps for a sample image.


2.4 Loss Function

Our loss function is a weighted sum of rate and distortion,

, as shown in (4). The parameter controls the trade-off between the rate loss and the distortion loss. We estimate the rate loss by summing up all values of the importance maps, . For distortion, we calculate both the Mean Squared Error (MSE) and the Multi-Scale Structural Similarity (MS-SSIM) index between the original image and the reconstructed image as in (5), where both and are trained to minimize (4).


2.5 Entropy Coding

Figure 3: Binary arithmetic coding of feature maps.
1:Initialize final feature maps Z
2:for each feature maps do
3:    Initialize Block Flag to 0
4:    Initialize Sample Flag to 0
5:    Initialize All Zero Bitplane Flag to 1
6:    for each bitplane = MSB to LSB do
7:       if  then
8:          Set as per all zero bitplane state
9:          Encode        
10:       if  then
11:          for each block  do
12:              if  then
13:                 Set as per significant block state
14:                 Encode               
15:              if  then
16:                 for each sample  do
17:                    if  then
18:                        Set as per significance state
19:                        Encode
20:                    else
21:                        Encode                                                                         
Algorithm 1 Binary arithmetic coding of feature maps

Our entropy coding is inspired by the JPEG2000 standard. It encodes the binary representation of individual final feature maps from the most significant bitplane (MSB) to the least significant one (LSB). Furthermore, on each bitplane, the coding bits are classified into the significant and refinement types, and coded on a block by block basis through a context-adaptive binary arithmetic coder. The parameter for block size is experimental. As with JPEG2000, separate context models are used for different types of bit, with the context probabilities updated constantly along the way. Additionally, we introduce flags at both the bitplane and block levels to indicate the all zero cases. More details are presented in Algorithm 1.

The entropy decoder decodes the compressed bitstream and recover the binary feature maps. The encoded bitstream also contains side information needed by the decoder. Figure 3 illustrates the procedure of our entropy coding.

3 Training

This section describes our training procedure and parameter settings.

3.1 Datasets

Our training dataset contains 10,000 images from the ImageNet Large Scale Visual Recognition Competition (ILSVRC) 2012

[4]. For training, we randomly crop 128x128 patches from these images as inputs to the autoencoder. For testing and validation, we use Kodak PhotoCD dataset.

3.2 Procedure

The training for the autoencoder and importance net is done in three sequential phases. In the first pre-training phase, the autoencoder is trained independently without the importance net. This is achieved by minimizing the reconstruction distortion alone. In the second phase, the importance net is included in the training process, to minimize an objective function that involves both the distortion and rate estimate. In this phase, however, the pre-trained autoencoder is not updated. In the final fine-tuning phase, both the autoencoder and the importance net are updated jointly. Table 1 shows the detailed parameter settings.

Parameters Values
[0.001, 10]
Batch size 10
Bit depth () 4
Block size 4 4
Learning rate [1e-4, 1e-6]
Channel number 16
Table 1: Parameter settings.

4 Evaluation

This section compares the compression performance of our system with that of JPEG and BPG at  0.15 bpp in terms of PSNR and MS-SSIM. The test dataset is from CLIC 2018. The results are presented in Table 2.

JPEG 0.155 22.1 0.7568
BPG 0.153 29.2 0.9425
Ours 0.126 26.3 0.9242
Table 2: Compression performance comparison.

It is worth noting that the PSNR numbers are computed following the CLIC 2018 method; that is, the mean squared error over these 287 test images is first calculated before the PSNR conversion. For MS-SSIM, we acquire the numbers by averaging the MS-SSIM values of these test images, with the MS-SSIM of each test image obtained by taking the average over the three color components. From Table 2, both BPG and our system outperform JPEG by a large margin in terms of PSNR. Although the BPG is 3dB better than ours in terms of PSNR, their MS-SSIM are close to each other. Figure 4 presents few sample images to compare the subjective quality of these methods.

(a) Original image (b) JPEG
(c) BPG (d) Our System
Figure 4: Original image (a) and the reconstructed images produced by (b) JPEG , (c) BPG, and (d) our system.

5 Conclusion

This work describes NCTU’s response to the CLIC 2018. Its presents an autoencoder-based learned image compressor with the notion of importance maps for bit allocation and rate control. While it outperforms JPEG significantly, there is still a performance gap relative to BPG, suggesting ample room for further improvements. Among others, the impact of importance maps on subjective quality deserves further investigation. In our case, it causes color banding in several images.

6 Acknowledgement

This work was supported in part by the MOST, Taiwan under Grant MOST 107-2634-F-009-007. The authors are grateful to the National Center for High Performance Computing, Taiwan, for providing GPU facility. Also, we like to thank Ms. Yang-Tzu Liu Tsen for her help on the computer simulation and arithmetic coder implementation.