-
TCIM: Triangle Counting Acceleration With Processing-In-MRAM Architecture
Triangle counting (TC) is a fundamental problem in graph analysis and ha...
read it
-
Hyperdimensional Computing Nanosystem
One viable solution for continuous reduction in energy-per-operation is ...
read it
-
A Memristor based Unsupervised Neuromorphic System Towards Fast and Energy-Efficient GAN
Deep Learning has gained immense success in pushing today's artificial i...
read it
-
CPU and GPU Accelerated Fully Homomorphic Encryption
Fully Homomorphic Encryption (FHE) is one of the most promising technolo...
read it
-
LSTM-Sharp: An Adaptable, Energy-Efficient Hardware Accelerator for Long Short-Term Memory
The effectiveness of LSTM neural networks for popular tasks such as Auto...
read it
-
Fast Arithmetic Hardware Library For RLWE-Based Homomorphic Encryption
In this work, we propose an open-source, first-of-its-kind, arithmetic h...
read it
-
Neurostream: Scalable and Energy Efficient Deep Learning with Smart Memory Cubes
High-performance computing systems are moving towards 2.5D and 3D memory...
read it
Computing-in-Memory for Performance and Energy Efficient Homomorphic Encryption
Homomorphic encryption (HE) allows direct computations on encrypted data. Despite numerous research efforts, the practicality of HE schemes remains to be demonstrated. In this regard, the enormous size of ciphertexts involved in HE computations degrades computational efficiency. Near-memory Processing (NMP) and Computing-in-memory (CiM) - paradigms where computation is done within the memory boundaries - represent architectural solutions for reducing latency and energy associated with data transfers in data-intensive applications such as HE. This paper introduces CiM-HE, a Computing-in-memory (CiM) architecture that can support operations for the B/FV scheme, a somewhat homomorphic encryption scheme for general computation. CiM-HE hardware consists of customized peripherals such as sense amplifiers, adders, bit-shifters, and sequencing circuits. The peripherals are based on CMOS technology, and could support computations with memory cells of different technologies. Circuit-level simulations are used to evaluate our CiM-HE framework assuming a 6T-SRAM memory. We compare our CiM-HE implementation against (i) two optimized CPU HE implementations, and (ii) an FPGA-based HE accelerator implementation. When compared to a CPU solution, CiM-HE obtains speedups between 4.6x and 9.1x, and energy savings between 266.4x and 532.8x for homomorphic multiplications (the most expensive HE operation). Also, a set of four end-to-end tasks, i.e., mean, variance, linear regression, and inference are up to 1.1x, 7.7x, 7.1x, and 7.5x faster (and 301.1x, 404.6x, 532.3x, and 532.8x more energy efficient). Compared to CPU-based HE in a previous work, CiM-HE obtain 14.3x speed-up and >2600x energy savings. Finally, our design offers 2.2x speed-up with 88.1x energy savings compared to a state-of-the-art FPGA-based accelerator.
READ FULL TEXT
Comments
There are no comments yet.