Distributed-Shared Memory Computed Tomography

被引:0
|
作者
de la Fuente, Francisco [1 ]
Torres, Felipe [1 ]
Rannou, Fernando R. [1 ]
机构
[1] Univ Santiago Chile, Dept Ingn Informat, Av Ecuador 3659, Santiago, Chile
关键词
Distributed-shared memory; Statistical Reconstruction;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Large-scale statistical reconstruction algorithms are known to be memory and processor intensive applications. For instance, the system matrix for a small animal scanner requires several gigabytes of memory storage and the algorithm usually needs many iterations to produce acceptable images. In this abstract we design distributed-shared memory (DSM) statistical reconstruction algorithms to exploit all available computational resources as a unified infrastructure and thereby improving the cost-efficiency of the investment and scalability of the system. We use and compare two distinct approaches. The first one uses the Unified Parallel C (UPC) compiler which transparently provides a global shared virtual address space across all computers. Data is physically stored in different computers, but threads can access any shared item as it if were in its local memory. The second approach combines OpenMP and Pthreads shared-memory libraries with the message-passing library MPI. In this case threads only have access to the node's local memory and access to remote data is carried out explicitly through message-passing. Early UPC experiments showed that keeping all data shared heavily affects reconstruction performance. Therefore, we devised a distribution method where some data is kept shared and other is kept private, mimicking somehow the library-based approach. However, even with data privatization, the compiler solution cannot compete with the library solutions. We explore three workload distribution strategies: LOR-based, Nonzero-based and Cores-based. The best performance is obtained with OpenMP+MPI and the Core-based balance algorithm, which reaches a speedup of 36 with 112 cores. However, both OpenMP+MPI and Pthreads+MPI outperform UPC by large. The low system efficiency of 0.32 is mainly due to the slow internode communication network.
引用
收藏
页码:2452 / 2455
页数:4
相关论文
共 50 条
  • [31] An open Distributed Shared Memory system
    Manis, G
    Lymberopoulos, L
    Koziris, N
    Papakonstantinou, G
    HIGH-PERFORMANCE COMPUTING AND NETWORKING, 2001, 2110 : 293 - 301
  • [32] ALGORITHMS IMPLEMENTING DISTRIBUTED SHARED MEMORY
    STUMM, M
    ZHOU, SN
    COMPUTER, 1990, 23 (05) : 54 - 64
  • [33] Debugging distributed shared memory applications
    Olivier, Jeffrey
    Chen, Chih-Ping
    Hoeflinger, Jay
    PARALLEL AND DISTRIBUTED PROCESSING AND APPLICATIONS, 2006, 4330 : 833 - +
  • [34] A transparent Software Distributed Shared Memory
    Kohn, ED
    Schuster, A
    EURO-PAR 2003 PARALLEL PROCESSING, PROCEEDINGS, 2003, 2790 : 1180 - 1189
  • [35] AN ANALYSIS OF DISTRIBUTED SHARED MEMORY ALGORITHMS
    KESSLER, RE
    LIVNY, M
    9TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS, 1989, : 498 - 505
  • [36] A distributed shared memory programming course
    Wilkinson, B
    Pai, T
    Miraj, M
    FIRST IEEE/ACM INTERNATIONAL SYMPOSIUM ON CLUSTER COMPUTING AND THE GRID, PROCEEDINGS, 2001, : 492 - 497
  • [37] A UNIX distributed shared memory system
    Mendoza, R
    Gutierrez, ED
    Williams, DH
    INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED PROCESSING TECHNIQUES AND APPLICATIONS, VOLS I-V, PROCEEDINGS, 1999, : 713 - 717
  • [38] COMPARING SHARED AND DISTRIBUTED MEMORY COMPUTERS
    BAILLIE, CF
    PARALLEL COMPUTING, 1988, 8 (1-3) : 101 - 110
  • [39] Constructive and adaptable distributed shared memory
    Bataller, J
    Bernabeu-Auban, JM
    THIRD INTERNATIONAL WORKSHOP ON HIGH-LEVEL PARALLEL PROGRAMMING MODELS AND SUPPORTIVE ENVIRONMENTS, PROCEEDINGS, 1998, : 15 - 22
  • [40] Software Transactional Distributed Shared Memory
    Dash, Alokika
    Demsky, Brian
    ACM SIGPLAN NOTICES, 2009, 44 (04) : 297 - 298