In-Memory Performance for Big Data

被引:30
|
作者
Graefe, Goetz [1 ]
Volos, Haris [1 ]
Kimura, Hideaki [1 ]
Kuno, Harumi [1 ]
Tucek, Joseph [1 ]
Lillibridge, Mark [1 ]
Veitch, Alistair [1 ]
机构
[1] HP Labs Palo Alto, Palo Alto, CA 94304 USA
来源
PROCEEDINGS OF THE VLDB ENDOWMENT | 2014年 / 8卷 / 01期
关键词
D O I
10.14778/2735461.2735465
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
When a working set fits into memory, the overhead imposed by the buffer pool renders traditional databases noncompetitive with in-memory designs that sacrifice the benefits of a buffer pool. However, despite the large memory available with modern hardware, data skew, shifting workloads, and complex mixed workloads make it difficult to guarantee that a working set will fit in memory. Hence, some recent work has focused on enabling in-memory databases to protect performance when the working data set almost fits in memory. Contrary to those prior efforts, we enable buffer pool designs to match in-memory performance while supporting the "big data" workloads that continue to require secondary storage, thus providing the best of both worlds. We introduce here a novel buffer pool design that adapts pointer swizzling for references between system objects (as opposed to application objects), and uses it to practically eliminate buffer pool overheads for memory resident data. Our implementation and experimental evaluation demonstrate that we achieve graceful performance degradation when the working set grows to exceed the buffer pool size, and graceful improvement when the working set shrinks towards and below the memory and buffer pool sizes.
引用
收藏
页码:37 / 48
页数:12
相关论文
共 50 条
  • [41] Enhancing the Interactive Visualisation of a Data Preparation Tool from in-Memory Fitting to Big Data Sets
    Epelde, Gorka
    Alvarez, Roberto
    Beristain, Andoni
    Arrue, Monica
    Arangoa, Itsasne
    Rankin, Debbie
    [J]. BUSINESS INFORMATION SYSTEMS WORKSHOPS (BIS 2020), 2020, 394 : 272 - 284
  • [42] Data Prefetching and Eviction Mechanisms of In-Memory Storage Systems Based on Scheduling for Big Data Processing
    Chen, Chien-Hung
    Hsia, Ting-Yuan
    Huang, Yennun
    Kuo, Sy-Yen
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2019, 30 (08) : 1738 - 1752
  • [43] Performance enhancement for iterative data computing with in-memory concurrent processing
    Wen, Yean-Fu
    Chen, Yu-Fang
    Chiu, Tse Kai
    Chen, Yen-Chou
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2020, 32 (07):
  • [44] Performance Analysis of a Multi-Tenant In-memory Data Grid
    Das, Anwesha
    Mueller, Frank
    Gu, Xiaohui
    Iyengar, Arun
    [J]. PROCEEDINGS OF 2016 IEEE 9TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING (CLOUD), 2016, : 956 - 959
  • [45] Performance Characterization of In-Memory Data Analytics on a Modern Cloud Server
    Awan, Ahsan Javed
    Brorsson, Mats
    Vlassov, Vladimir
    Ayguade, Eduard
    [J]. PROCEEDINGS 2015 IEEE FIFTH INTERNATIONAL CONFERENCE ON BIG DATA AND CLOUD COMPUTING BDCLOUD 2015, 2015, : 1 - 8
  • [46] In-Memory Big Graph: A Future Research Agenda
    Jain, Deepali
    Patgiri, Ripon
    Nayak, Sabuzima
    [J]. BUSINESS INFORMATION SYSTEMS, PT I, 2019, 353 : 18 - 29
  • [47] In-Memory Data Flow Processor
    Fujiki, Daichi
    Mahlke, Scott
    Das, Reetuparna
    [J]. 2017 26TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT), 2017, : 375 - 375
  • [48] An In-Memory Data-Cube Aware Distributed Data Discovery Across Clouds for Remote Sensing Big Data
    Song, Jie
    Ma, Yan
    Zhang, Zhixin
    Liu, Peng
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2023, 16 : 4529 - 4548
  • [49] In-Memory Data Parallel Processor
    Fujiki, Daichi
    Mahlke, Scott
    Das, Reetuparna
    [J]. ACM SIGPLAN NOTICES, 2018, 53 (02) : 1 - 14
  • [50] GFlink: An In-Memory Computing Architecture on Heterogeneous CPU-GPU Clusters for Big Data
    Chen, Cen
    Li, Kenli
    Ouyang, Aijia
    Tang, Zhuo
    Li, Keqin
    [J]. PROCEEDINGS 45TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING - ICPP 2016, 2016, : 542 - 551