Low-Rank Winograd Transformation for 3D Convolutional Neural Networks

01/26/2023
by   Ziran Qin, et al.
0

This paper focuses on Winograd transformation in 3D convolutional neural networks (CNNs) that are more over-parameterized compared with the 2D version. The over-increasing Winograd parameters not only exacerbate training complexity but also barricade the practical speedups due simply to the volume of element-wise products in the Winograd domain. We attempt to reduce trainable parameters by introducing a low-rank Winograd transformation, a novel training paradigm that decouples the original large tensor into two less storage-required trainable tensors, leading to a significant complexity reduction. Built upon our low-rank Winograd transformation, we take one step ahead by proposing a low-rank oriented sparse granularity that measures column-wise parameter importance. By simply involving the non-zero columns in the element-wise product, our sparse granularity is empowered with the ability to produce a very regular sparse pattern to acquire effectual Winograd speedups. To better understand the efficacy of our method, we perform extensive experiments on 3D CNNs. Results manifest that our low-rank Winograd transformation well outperforms the vanilla Winograd transformation. We also show that our proposed low-rank oriented sparse granularity permits practical Winograd acceleration compared with the vanilla counterpart.

READ FULL TEXT
research
05/24/2019

Learning Low-Rank Approximation for CNNs

Low-rank approximation is an effective model compression technique to no...
research
02/07/2017

Low Rank Matrix Recovery with Simultaneous Presence of Outliers and Sparse Corruption

We study a data model in which the data matrix D can be expressed as D =...
research
10/31/2018

Low-Rank Embedding of Kernels in Convolutional Neural Networks under Random Shuffling

Although the convolutional neural networks (CNNs) have become popular fo...
research
06/01/2018

IGCV3: Interleaved Low-Rank Group Convolutions for Efficient Deep Neural Networks

In this paper, we are interested in building lightweight and efficient c...
research
12/15/2010

TILT: Transform Invariant Low-rank Textures

In this paper, we show how to efficiently and effectively extract a clas...
research
08/12/2022

Parallel QR Factorization of Block Low-Rank Matrices

We present two new algorithms for Householder QR factorization of Block ...
research
02/11/2017

Group Scissor: Scaling Neuromorphic Computing Design to Large Neural Networks

Synapse crossbar is an elementary structure in Neuromorphic Computing Sy...

Please sign up or login with your details

Forgot password? Click here to reset