Towards Continuous Benchmarking: An Automated Performance Evaluation Framework for High Performance Software

被引:3
|
作者
Anzt, Hartwig [1 ,2 ]
Chen, Yen-Chen [3 ]
Cojean, Terry [1 ]
Dongarra, Jack [2 ,4 ,5 ]
Flegar, Goran [6 ]
Nayak, Pratik [1 ]
Quintana-Orti, Enrique S. [7 ]
Tsai, Yuhsiang M. [3 ]
Wang, Weichung [3 ]
机构
[1] Karlsruhe Inst Technol, Karlsruhe, Germany
[2] Univ Tennessee, Knoxville, TN 37996 USA
[3] Taiwan Natl Univ, Taipei, Taiwan
[4] Oak Ridge Natl Lab, Oak Ridge, TN USA
[5] Univ Manchester, Manchester, Lancs, England
[6] Univ Jaime I, Castellon de La Plana, Castello, Spain
[7] Univ Politecn Valencia, Valencia, Spain
关键词
interactive performance visualization; automated performance benchmarking; continuous integration; healthy software lifecycle;
D O I
10.1145/3324989.3325719
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
We present an automated performance evaluation framework that enables an automated workflow for testing and performance evaluation of software libraries. Integrating this component into an ecosystem enables sustainable software development, as a community effort, via a web application for interactively evaluating the performance of individual software components. The performance evaluation tool is based exclusively on web technologies, which removes the burden of downloading performance data or installing additional software. We employ this framework for the GINKGO software ecosystem, but the framework can be used with essentially any software project, including the comparison between different software libraries. The Continuous Integration (CI) framework of GINKGO is also extended to automatically run a benchmark suite on predetermined HPC systems, store the state of the machine and the environment along with the compiled binaries, and collect results in a publicly accessible performance data repository based on Git. The GINKGO performance explorer (GPE) can be used to retrieve the performance data from the repository, and visualizes it in a web browser. GPE also implements an interface that allows users to write scripts, archived in a Git repository, to extract particular data, compute particular metrics, and visualize them in many different formats (as specified by the script). The combination of these approaches creates a workflow which enables performance reproducibility and software sustainability of scientific software. In this paper, we present example scripts that extract and visualize performance data for GINKGO'S SpMV kernels that allow users to identify the optimal kernel for specific problem characteristics.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Towards a framework of the performance evaluation of SMEs' industry portals
    Chou, TC
    Hsu, LL
    Yeh, YJ
    Ho, CT
    [J]. INDUSTRIAL MANAGEMENT & DATA SYSTEMS, 2005, 105 (3-4) : 527 - 544
  • [32] A Framework for the Relative and Absolute Performance Evaluation of Automated Spectroscopy Systems
    Portnoy, David
    Heimberg, Peter
    Heimberg, Jennifer
    Feuerbach, Robert
    McQuarrie, Allan
    Noonan, William
    Mattson, John
    [J]. INTERNATIONAL CONFERENCE ON APPLICATIONS OF NUCLEAR TECHNIQUES, 2009, 1194 : 145 - 159
  • [33] AutoDFEK: Automated Declarative Performance Evaluation and Tuning Framework on Kubemetes
    Choochotkaew, Sunyanan
    Chiba, Tatsuhiro
    Trent, Scott
    Yoshimura, Takeshi
    Amaral, Marcelo
    [J]. 2022 IEEE 15TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING (IEEE CLOUD 2022), 2022, : 309 - 314
  • [34] Benchmarking human performance for continuous speech recognition
    Deshmukh, N
    Duncan, RJ
    Ganapathiraju, A
    Picone, J
    [J]. ICSLP 96 - FOURTH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, VOLS 1-4, 1996, : 2486 - 2489
  • [35] Using Benchmarking Bots for Continuous Performance Assessment
    Markusse, Florian
    Leitner, Philipp
    Serebrenik, Alexander
    [J]. IEEE SOFTWARE, 2022, 39 (05) : 50 - 55
  • [36] A benchmarking framework for understanding bus performance in the US
    Morse, Lindsey
    Trompet, Mark
    Barron, Alexander
    Anderson, Richard
    Graham, Daniel J.
    [J]. BENCHMARKING-AN INTERNATIONAL JOURNAL, 2020, 27 (04) : 1533 - 1550
  • [37] Towards high performance security policy evaluation
    Zheng Qin
    Fei Chen
    Qiang Wang
    Alex X. Liu
    Zhiguang Qin
    [J]. The Journal of Supercomputing, 2012, 59 : 1577 - 1595
  • [38] Towards high performance security policy evaluation
    Qin, Zheng
    Chen, Fei
    Wang, Qiang
    Liu, Alex X.
    Qin, Zhiguang
    [J]. JOURNAL OF SUPERCOMPUTING, 2012, 59 (03): : 1577 - 1595
  • [39] Democratizing digital design and manufacturing using high performance cloud computing: Performance evaluation and benchmarking
    Wu, Dazhong
    Liu, Xi
    Hebert, Steve
    Gentzsch, Wolfgang
    Terpenny, Janis
    [J]. JOURNAL OF MANUFACTURING SYSTEMS, 2017, 43 : 316 - 326
  • [40] Towards an automated evaluation process for software architectures
    Bashroush, R
    Spence, I
    Kilpatrick, P
    Brown, TJ
    [J]. PROCEEDINGS OF THE IASTED INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, 2004, : 54 - 58