HashMem: PIM-based Hashmap Accelerator

06/30/2023
by   Akhil Shekar, et al.
0

Hashmaps are widely utilized data structures in many applications to perform a probe on key-value pairs. However, their performance tends to degrade with the increase in the dataset size, which leads to expensive off-chip memory accesses to perform bucket traversals associated with hash collision. In this work, we propose HashMem, a processing-in-memory (PIM) architecture designed to perform bucket traversals along the row buffers at the subarray level. Due to the inherent parallelism achieved with many concurrent subarray accesses and the massive bandwidth available within DRAM, the execution time related to bucket traversals is significantly reduced. We have evaluated two versions of HashMem, performance-optimized and area-optimized, which have a speedup of 49.1x/17.1x and 9.2x/3.2x over standard C++ map and hyper-optimized hopscotch map implementations, respectively.

READ FULL TEXT
research
04/15/2021

pLUTo: In-DRAM Lookup Tables to Enable Massively Parallel General-Purpose Computation

Data movement between main memory and the processor is a significant con...
research
09/21/2022

HiRA: Hidden Row Activation for Reducing Refresh Latency of Off-the-Shelf DRAM Chips

DRAM is the building block of modern main memory systems. DRAM cells mus...
research
11/03/2020

Booster: An Accelerator for Gradient Boosting Decision Trees

We propose Booster, a novel accelerator for gradient boosting trees base...
research
08/18/2022

GRAPHIC: GatheR-And-Process in Highly parallel with In-SSD Compression Architecture in Very Large-Scale Graph

Graph convolutional network (GCN), an emerging algorithm for graph compu...
research
08/01/2018

MARS: Memory Aware Reordered Source

Memory bandwidth is critical in today's high performance computing syste...
research
02/15/2021

GradPIM: A Practical Processing-in-DRAM Architecture for Gradient Descent

In this paper, we present GradPIM, a processing-in-memory architecture w...
research
07/09/2019

Depth from Small Motion using Rank-1 Initialization

Depth from Small Motion (DfSM) (Ha et al., 2016) is particularly interes...

Please sign up or login with your details

Forgot password? Click here to reset