MSIM : A Highly Parallel Near-Memory Accelerator for MinHash Sketch

被引:0
|
作者
Sinha, Aman [1 ]
Mai, Jhih-Yong [1 ]
Lai, Bo-Cheng [1 ]
机构
[1] Natl Yang Ming Chiao Tung Univ, Inst Elect, Hsinchu, Taiwan
关键词
Processing-In-Memory; Near Memory Processing; MinHash Sketches; Long read genome assembly;
D O I
10.1109/SOCC56010.2022.9908115
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Genome Assembly is an important Big Data analytics which involves massive computations for similarity searches on sequence databases. Being major component of runtime, similarity searches require careful design for scalable performance. MinHash Sketching is an extensively used data structure in Long-read genome assembly pipelines, which involves generating, randomizing and minimizing a set of hashes for all the k-mers in genome sequences. Compute-hungry MinHash sketch processing on commercially available multi-threaded CPUs suffer from the limited bandwidth of the L1-cache, which causes the CPUs to stall. Near-Data Processing (NDP) is an emerging trend in data-bound Big Data analytics to harness the low-latency, high-bandwidth available within the Dual In-line Memory Modules (DIMMs). While NDP architectures have generally been utilized for memory-bound computations, MinHash sketching is a potential application that can gain massive throughput by exploiting memory Banks as higher bandwidth L1-cache. In this work, we propose MSIM, a distributed, highly parallel and efficient hardware-software co-design for accelerating MinHash Sketch processing on light-weight components placed on the DRAM hierarchy. Multiple ASIC-based Processing Engines (PEs) placed at the bank-group-level in MSIM provide high-parallelism for low-latency computations. The PEs sequentially access data from all Banks within their bank-group with the help of a dedicated Address calculator, which utilizes an optimal data mapping scheme. The PEs are controlled by a custom Arbiter, which is directly activated by the host CPU using general DDR commands, without requiring any modification to the memory controller or the DIMM standard buses. MSIM requires limited area and power overheads, while displaying up-to 384.9x speedup and 1088.4x energy reduction compared to the baseline multi-threaded software solution in our experiments. MSIM achieves 4.26x speedup over high-end GPU, while consuming 26.4x lesser energy. Moreover, MSIM design is highly scalable and extendable in nature.
引用
收藏
页码:184 / 189
页数:6
相关论文
共 50 条
  • [21] Near-memory computing: Past, present, and future
    Singh, Gagandeep
    Chelini, Lorenzo
    Corda, Stefano
    Awan, Ahsan Javed
    Stuijk, Sander
    Jordans, Roel
    Corporaal, Henk
    Boonstra, Albert-Jan
    MICROPROCESSORS AND MICROSYSTEMS, 2019, 71
  • [22] Concurrent Data Structures for Near-Memory Computing
    Liu, Zhiyu
    Calciu, Irina
    Herlihy, Maurice
    Mutlu, Onur
    PROCEEDINGS OF THE 29TH ACM SYMPOSIUM ON PARALLELISM IN ALGORITHMS AND ARCHITECTURES (SPAA'17), 2017, : 235 - 245
  • [23] A 93 TOPS/Watt Near-Memory Reconfigurable SAD Accelerator for HEVC/AV1/JEM Encoding
    Sundaram, Jainaveen
    Srinivasa, Srivatsa Rangachar
    Kurian, Dileep
    Chakraborty, Indranil
    Kale, Sirisha
    Jain, Nilesh
    Karnik, Tanay
    Iyer, Ravi
    Srinivasan, Anuradha
    PROCEEDINGS OF THE 2021 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2021), 2021, : 1400 - 1403
  • [24] Scalable Phylogeny Reconstruction with Disaggregated Near-memory Processing
    Alachiotis, Nikolaos
    Skrimponis, Panagiotis
    Pissadakis, Manolis
    Pnevmatikatos, Dionisios
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2022, 15 (03)
  • [25] Operand-Oriented Virtual Memory Support for Near-Memory Processing
    Choi, Duheon
    Jeong, Taeyang
    Yeom, Joonhyeok
    Chung, Eui-Young
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (08) : 2250 - 2263
  • [26] HMComp: Extending Near-Memory Capacity using Compression in Hybrid Memory
    Shao, Qi
    Arelakis, Angelos
    Stenstrom, Per
    PROCEEDINGS OF THE 38TH ACM INTERNATIONAL CONFERENCE ON SUPERCOMPUTING, ACM ICS 2024, 2024, : 74 - 84
  • [27] Guest Editors' Introduction: Near-Memory and In-Memory Processing Introduction
    Li, Hai Helen
    Alameldeen, Alaa R.
    Mutlu, Onur
    IEEE DESIGN & TEST, 2022, 39 (02) : 46 - 47
  • [28] A Review of Near-Memory Computing Architectures: Opportunities and Challenges
    Singh, Gagandeep
    Chelini, Lorenzo
    Corda, Stefano
    Awan, Ahsan Javed
    Stuijk, Sander
    Jordans, Roel
    Corporaal, Henk
    Boonstra, Albert-Jan
    2018 21ST EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD 2018), 2018, : 608 - 617
  • [29] Challenges and Opportunities: From Near-memory Computing to In-memory Computing
    Khoram, Soroosh
    Zha, Yue
    Zhang, Jialiang
    Li, Jing
    ISPD'17: PROCEEDINGS OF THE 2017 ACM INTERNATIONAL SYMPOSIUM ON PHYSICAL DESIGN, 2017, : 43 - 46
  • [30] ACTS: A Near-Memory FPGA Graph Processing Framework
    Jaiyeoba, Wole
    Elyasi, Nima
    Choi, Changho
    Skadron, Kevin
    PROCEEDINGS OF THE 2023 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD PROGRAMMABLE GATE ARRAYS, FPGA 2023, 2023, : 79 - 89