Memory Debugging of MPI-Parallel Applications in Open MPI

被引:0
|
作者
Keller, Rainer [1 ]
Fan, Shiqing [1 ]
Resch, Michael [1 ]
机构
[1] Univ Stuttgart, Ctr High Performance Comp, D-7000 Stuttgart, Germany
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In this paper we describe the implementation of memory checking functionality based on instrumentation using valgrind. The combination of valgrind based checking functions within the MPI-implementation offers superior debugging functionality, for errors that otherwise are not possible to detect with comparable MPI-debugging tools. The functionality is integrated into Open MPI as the so-called memchecker-framework. This allows other memory debuggers that offer a similar API to be integrated. The tight control of the user's memory passed to Open MPI, allows not only to find application errors. but also helps track bugs within Open MPI itself. We describe the actual checks, classes of errors being found, how memory buffers internally are being handled, show errors actually found in user's code and the performance implications of this instrumentation.
引用
下载
收藏
页码:517 / 523
页数:7
相关论文
共 50 条
  • [41] Concurrent programming constructs for parallel MPI applicationsThe MPI threads library
    Tobias Berka
    Giorgos Kollias
    Helge Hagenauer
    Marian Vajteršic
    Ananth Grama
    The Journal of Supercomputing, 2013, 63 : 385 - 406
  • [42] INTERPRETIVE MPI FOR PARALLEL COMPUTING
    Chou, Yu-Cheng
    Cheng, Harry H.
    DETC 2008: PROCEEDINGS OF THE ASME INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATIONAL IN ENGINEERING CONFERENCE, VOL 3, PTS A AND B: 28TH COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, 2009, : 1163 - 1176
  • [43] Diagnostic methods for communication waiting in MPI parallel programs and applications
    Wu L.
    Jing C.
    Liu X.
    Tian H.
    Guofang Keji Daxue Xuebao/Journal of National University of Defense Technology, 2020, 42 (02): : 47 - 54
  • [44] Fault tolerance for cluster-oriented MPI parallel applications
    Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China
    Qinghua Daxue Xuebao, 2006, 1 (67-69+110):
  • [45] FIMD-MPI: A tool for injecting faults into MPI applications
    Blough, Douglas M.
    Liu, Peng
    Proceedings of the International Parallel Processing Symposium, IPPS, 2000, : 241 - 247
  • [46] MPI Based Cluster Computing for Performance Evaluation of Parallel Applications
    Nanjesh, B. R.
    Kumar, Vinay K. S.
    Madhu, C. K.
    Kumar, Hareesh G.
    2013 IEEE CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGIES (ICT 2013), 2013, : 1123 - 1128
  • [47] Automatic Partitioning of MPI Operations in MPI plus OpenMP Applications
    Jammer, Tim
    Bischof, Christian
    HIGH PERFORMANCE COMPUTING - ISC HIGH PERFORMANCE DIGITAL 2021 INTERNATIONAL WORKSHOPS, 2021, 12761 : 191 - 198
  • [48] MPI Stages: Checkpointing MPI State for Bulk Synchronous Applications
    Sultana, Nawrin
    Skjellum, Anthony
    Laguna, Ignacio
    Farmer, Matthew Shane
    Mohror, Kathryn
    Emani, Murali
    EUROMPI 2018: PROCEEDINGS OF THE 25TH EUROPEAN MPI USERS' GROUP MEETING, 2018,
  • [49] MPI Collectives for Multi-core Clusters: Optimized Performance of the Hybrid MPI plus MPI Parallel Codes
    Zhou, Huan
    Gracia, Jose
    Schneider, Ralf
    PROCEEDINGS OF THE 48TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING WORKSHOPS (ICPP 2019), 2019,
  • [50] MPC-MPI: An MPI Implementation Reducing the Overall Memory Consumption
    Perache, Marc
    Carribault, Patrick
    Jourdren, Herve
    RECENT ADVANCES IN PARALLEL VIRTUAL MACHINE AND MESSAGE PASSING INTERFACE, PROCEEDINGS, 2009, 5759 : 94 - 103