Compressed Learning of Deep Neural Networks for OpenCL-Capable Embedded Systems

05/20/2019
by   Sangkyun Lee, et al.
0

Deep neural networks (DNNs) have been quite successful in solving many complex learning problems. However, DNNs tend to have a large number of learning parameters, leading to a large memory and computation requirement. In this paper, we propose a model compression framework for efficient training and inference of deep neural networks on embedded systems. Our framework provides data structures and kernels for OpenCL-based parallel forward and backward computation in a compressed form. In particular, our method learns sparse representations of parameters using ℓ_1-based sparse coding while training, storing them in compressed sparse matrices. Unlike the previous works, our method does not require a pre-trained model as an input and therefore can be more versatile for different application environments. Even though the use of ℓ_1-based sparse coding for model compression is not new, we show that it can be far more effective than previously reported when we use proximal point algorithms and the technique of debiasing. Our experiments show that our method can produce minimal learning models suitable for small embedded devices.

READ FULL TEXT

page 2

page 3

page 10

page 12

page 16

page 22

page 23

page 24

research
11/16/2021

Neuron-based Pruning of Deep Neural Networks with Better Generalization using Kronecker Factored Curvature Approximation

Existing methods of pruning deep neural networks focus on removing unnec...
research
05/27/2018

Compact and Computationally Efficient Representation of Deep Neural Networks

Dot product operations between matrices are at the heart of almost any f...
research
10/21/2018

To Compress, or Not to Compress: Characterizing Deep Learning Model Compression for Embedded Inference

The recent advances in deep neural networks (DNNs) make them attractive ...
research
11/01/2017

Efficient Inferencing of Compressed Deep Neural Networks

Large number of weights in deep neural networks makes the models difficu...
research
09/23/2019

Class-dependent Compression of Deep Neural Networks

Today's deep neural networks require substantial computation resources f...
research
11/10/2020

Neural Network Compression Via Sparse Optimization

The compression of deep neural networks (DNNs) to reduce inference cost ...
research
07/19/2021

A New Clustering-Based Technique for the Acceleration of Deep Convolutional Networks

Deep learning and especially the use of Deep Neural Networks (DNNs) prov...

Please sign up or login with your details

Forgot password? Click here to reset