Understanding the Behavior of In-Memory Computing Workloads

被引:0
|
作者
Jiang, Tao [1 ,2 ]
Zhang, Qianlong [1 ]
Hou, Rui [1 ]
Chai, Lin [1 ]
Mckee, Sally A. [3 ]
Jia, Zhen [1 ]
Sun, Ninghui [1 ]
机构
[1] Chinese Acad Sci, ICT, SKL Comp Architecture, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Chalmers Univ Technol, Gothenburg, Sweden
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The increasing demands of big data applications have led researchers and practitioners to turn to in-memory computing to speed processing. For instance, the Apache Spark framework stores intermediate results in memory to deliver good performance on iterative machine learning and interactive data analysis tasks. To the best of our knowledge, though, little work has been done to understand Spark's architectural and microarchitectural behaviors. Furthermore, although conventional commodity processors have been well optimized for traditional desktops and HPC, their effectiveness for Spark workloads remains to be studied. To shed some light on the effectiveness of conventional general-purpose processors on Spark workloads, we study their behavior in comparison to those of Hadoop, CloudSuite, SPEC CPU2006, TPC-C, and DesktopCloud. We evaluate the benchmarks on a 17-node Xeon cluster. Our performance results reveal that Spark workloads have significantly different characteristics from Hadoop and traditional HPC benchmarks. At the system level, Spark workloads have good memory bandwidth utilization (up to 50%), stable memory accesses, and high disk IO request frequency (200 per second). At the microarchitectural level, the cache and TLB are effective for Spark workloads, but the L2 cache miss rate is high. We hope this work yields insights for chip and datacenter system designers.
引用
收藏
页码:22 / 30
页数:9
相关论文
共 50 条
  • [1] COMPARING IMPLEMENTATIONS OF NEAR-DATA COMPUTING WITH IN-MEMORY MAPREDUCE WORKLOADS
    Pugsley, Seth H.
    Jestes, Jeffrey
    Balasubramonian, Rajeev
    Srinivasan, Vijayalakshmi
    Buyuktosunoglu, Alper
    Davis, Al
    Li, Feifei
    [J]. IEEE MICRO, 2014, 34 (04) : 44 - 52
  • [2] In-memory parallelism for database workloads
    Trancoso, P
    [J]. EURO-PAR 2002 PARALLEL PROCESSING, PROCEEDINGS, 2002, 2400 : 532 - 542
  • [3] Broad-Purpose In-Memory Computing for Signal Monitoring and Machine Learning Workloads
    Jain, Saurabh
    Lin, Longyang
    Alioto, Massimo
    [J]. IEEE SOLID-STATE CIRCUITS LETTERS, 2020, 3 : 394 - 397
  • [4] Scalable In-Memory Computing
    Uta, Alexandru
    Sandu, Andreea
    Costache, Stefania
    Kielmann, Thilo
    [J]. 2015 15TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND GRID COMPUTING, 2015, : 805 - 810
  • [5] Hyperspectral In-Memory Computing
    Latifpour, Mostafa Honari
    Park, Byoung Jun
    Yamamoto, Yoshihisa
    Suh, Myoung-Gyun
    [J]. 2024 OPTICAL FIBER COMMUNICATIONS CONFERENCE AND EXHIBITION, OFC, 2024,
  • [6] Computing In-Memory, Revisited
    Milojicic, Dejan
    Bresniker, Kirk
    Campbell, Gary
    Faraboschi, Paolo
    Strachan, John Paul
    Williams, Stan
    [J]. 2018 IEEE 38TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS), 2018, : 1300 - 1309
  • [7] In-memory computing with ferroelectrics
    Rui Yang
    [J]. Nature Electronics, 2020, 3 : 237 - 238
  • [8] In-memory mechanical computing
    Tie Mei
    Chang Qing Chen
    [J]. Nature Communications, 14
  • [9] In-memory hyperdimensional computing
    Karunaratne, Geethan
    Le Gallo, Manuel
    Cherubini, Giovanni
    Benini, Luca
    Rahimi, Abbas
    Sebastian, Abu
    [J]. NATURE ELECTRONICS, 2020, 3 (06) : 327 - +
  • [10] In-memory mechanical computing
    Mei, Tie
    Chen, Chang Qing
    [J]. NATURE COMMUNICATIONS, 2023, 14 (01)