A key measure of the performance of a distributed memory parallel program is the communication overhead. On most current parallel systems, sending data from a local to a remote processor still takes one or two orders of magnitude longer than the time to access data on a local processor. The behavior
Compiling data-parallel programs for clusters of SMPs
β Scribed by Siegfried Benkner; Thomas Brandes
- Publisher
- John Wiley and Sons
- Year
- 2004
- Tongue
- English
- Weight
- 233 KB
- Volume
- 16
- Category
- Article
- ISSN
- 1532-0626
- DOI
- 10.1002/cpe.767
No coin nor oath required. For personal study only.
π SIMILAR VOLUMES
We present Spar, a programming language for semi-automatic parallel programming, in particular for the programming of array-based applications. The language has been designed as part of the Automap project, in which a compiler and runtime system are being developed for distributed-memory systems. As
We describe a methodology for developing high performance programs running on clusters of SMP nodes. The SMP cluster programming methodology is based on a small prototype kernel (Simple) of collective communication primitives that make efficient use of the hybrid shared and message-passing environme