Performance and experience with LAPI-a new high-performance communication library for the IBM RS/6000 SP
- 27 November 2002
- conference paper
- Published by Institute of Electrical and Electronics Engineers (IEEE)
- No. 10637133,p. 260-266
- https://doi.org/10.1109/ipps.1998.669923
Abstract
LAPI is a low-level, high-performance communication interface available on the IBM RS/6000 SP system. It provides an active-message-like interface along with remote memory copy and synchronization functionality. It is designed primarily for use by experienced programmers in developing parallel subsystems, libraries and tools, but we also expect power programmers to use it in end-user applications. IBM developed LAPI as a part of a project with Pacific Northwest National Laboratory (PNNL) to optimize the performance of the Global Arrays (GA) toolkit and its applications on the IBM RS/6000 SP. We provide an overview of LAPI characteristics and discuss its differences from other models such as MPI-2. We present some base performance parameters of LAPI including latency and bandwidth and compare it with performance of the MPI/MPL. The Global Arrays library from PNNL was ported to LAPI to exploit the performance benefits of this new interface. Experience using LAPI to implement GA and the performance of the resulting library are presented.Keywords
This publication has 10 references indexed in Scilit:
- The Midway distributed shared memory systemPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2002
- Integrating polling, interrupts, and thread managementPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2002
- Global Arrays: a portable "shared-memory" programming model for distributed memory computersPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2002
- A massively parallel multireference configuration interaction program: The parallel COLUMBUS programJournal of Computational Chemistry, 1997
- Global arrays: A nonuniform memory access programming model for high-performance computersThe Journal of Supercomputing, 1996
- TreadMarks: shared memory computing on networks of workstationsComputer, 1996
- Approaches to large‐scale parallel self‐consistent field calculationsJournal of Computational Chemistry, 1995
- SP2 system architectureIBM Systems Journal, 1995
- Orca: a language for parallel programming of distributed systemsIEEE Transactions on Software Engineering, 1992
- Linda in contextCommunications of the ACM, 1989