Learning Low-Rank Approximation for CNNs

05/24/2019
by   Dongsoo Lee, et al.
0

Low-rank approximation is an effective model compression technique to not only reduce parameter storage requirements, but to also reduce computations. For convolutional neural networks (CNNs), however, well-known low-rank approximation methods, such as Tucker or CP decomposition, result in degraded model accuracy because decomposed layers hinder training convergence. In this paper, we propose a new training technique that finds a flat minimum in the view of low-rank approximation without a decomposed structure during training. By preserving the original model structure, 2-dimensional low-rank approximation demanding lowering (such as im2col) is available in our proposed scheme. We show that CNN models can be compressed by low-rank approximation with much higher compression ratio than conventional training methods while maintaining or even enhancing model accuracy. We also discuss various 2-dimensional low-rank approximation techniques for CNNs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/27/2022

CP decomposition and low-rank approximation of antisymmetric tensors

For the antisymmetric tensors the paper examines a low-rank approximatio...
research
11/19/2015

Convolutional neural networks with low-rank regularization

Large CNNs have delivered impressive performance in various computer vis...
research
05/03/2023

Multi-dimensional Signal Recovery using Low-rank Deconvolution

In this work we present Low-rank Deconvolution, a powerful framework for...
research
07/30/2018

Extreme Network Compression via Filter Group Approximation

In this paper we propose a novel decomposition method based on filter gr...
research
01/26/2023

Low-Rank Winograd Transformation for 3D Convolutional Neural Networks

This paper focuses on Winograd transformation in 3D convolutional neural...
research
05/14/2020

PENNI: Pruned Kernel Sharing for Efficient CNN Inference

Although state-of-the-art (SOTA) CNNs achieve outstanding performance on...
research
06/16/2021

Simultaneous Training of Partially Masked Neural Networks

For deploying deep learning models to lower end devices, it is necessary...

Please sign up or login with your details

Forgot password? Click here to reset