A Simple Method to Reduce Off-chip Memory Accesses on Convolutional Neural Networks

01/28/2019
by   Doyun Kim, et al.
0

For convolutional neural networks, a simple algorithm to reduce off-chip memory accesses is proposed by maximally utilizing on-chip memory in a neural process unit. Especially, the algorithm provides an effective way to process a module which consists of multiple branches and a merge layer. For Inception-V3 on Samsung's NPU in Exynos, our evaluation shows that the proposed algorithm makes off-chip memory accesses reduced by 1/50, and accordingly achieves 97.59 off-chip memory.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2020

Improving Memory Utilization in Convolutional Neural Network Accelerators

While the accuracy of convolutional neural networks has achieved vast im...
research
10/01/2021

Enhanced Multigradient Dilution Preparation

Abstract: In our paper the new algorithm enhanced multi gradient Dilutio...
research
07/14/2021

CNN-Cap: Effective Convolutional Neural Network Based Capacitance Models for Full-Chip Parasitic Extraction

Accurate capacitance extraction is becoming more important for designing...
research
02/15/2021

The Dependent Chip Model (DCM): a simple and more realistic alternative to the Independent Chip Model (ICM)

The Dependent Chip Model (DCM) is proposed as an alternative to the Inde...
research
12/10/2022

On coexistence of decentralized system (blockchain) and central management in Internet-of-Things

Networks are composed of logical nodes and edges for communications. The...
research
04/13/2021

Mitigating Adversarial Attack for Compute-in-Memory Accelerator Utilizing On-chip Finetune

Compute-in-memory (CIM) has been proposed to accelerate the convolution ...
research
06/27/2021

OCCAM: Optimal Data Reuse for Convolutional Neural Networks

Convolutional neural networks (CNNs) are emerging as powerful tools for ...

Please sign up or login with your details

Forgot password? Click here to reset