ELight: Enabling Efficient Photonic In-Memory Neurocomputing with Life Enhancement

12/15/2021
by   Hanqing Zhu, et al.
0

With the recent advances in optical phase change material (PCM), photonic in-memory neurocomputing has demonstrated its superiority in optical neural network (ONN) designs with near-zero static power consumption, time-of-light latency, and compact footprint. However, photonic tensor cores require massive hardware reuse to implement large matrix multiplication due to the limited single-core scale. The resultant large number of PCM writes leads to serious dynamic power and overwhelms the fragile PCM with limited write endurance. In this work, we propose a synergistic optimization framework, ELight, to minimize the overall write efforts for efficient and reliable optical in-memory neurocomputing. We first propose write-aware training to encourage the similarity among weight blocks, and combine it with a post-training optimization method to reduce programming efforts by eliminating redundant writes. Experiments show that ELight can achieve over 20X reduction in the total number of writes and dynamic power with comparable accuracy. With our ELight, photonic in-memory neurocomputing will step forward towards viable applications in machine learning with preserved accuracy, order-of-magnitude longer lifetime, and lower programming energy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/28/2023

Design Space Exploration for PCM-based Photonic Memory

The integration of silicon photonics (SiPh) and phase change materials (...
research
07/24/2021

Architecting Optically-Controlled Phase Change Memory

Phase Change Memory (PCM) is an attractive candidate for main memory as ...
research
05/31/2023

Integrated multi-operand optical neurons for scalable and hardware-efficient deep learning

The optical neural network (ONN) is a promising hardware platform for ne...
research
12/14/2021

Sherman: A Write-Optimized Distributed B+Tree Index on Disaggregated Memory

Memory disaggregation architecture physically separates CPU and memory i...
research
09/19/2023

Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity

With the fast growth of parameter size, it becomes increasingly challeng...
research
10/09/2020

A Survey of Non-Volatile Main Memory Technologies: State-of-the-Arts, Practices, and Future Directions

Non-Volatile Main Memories (NVMMs) have recently emerged as promising te...
research
05/18/2022

Single-Shot Optical Neural Network

As deep neural networks (DNNs) grow to solve increasingly complex proble...

Please sign up or login with your details

Forgot password? Click here to reset