High-level abstractions for message-passing parallel programming
β Scribed by Fan Chan; Jiannong Cao; Yudong Sun
- Publisher
- Elsevier Science
- Year
- 2003
- Tongue
- English
- Weight
- 684 KB
- Volume
- 29
- Category
- Article
- ISSN
- 0167-8191
No coin nor oath required. For personal study only.
β¦ Synopsis
Large-scale scientific and engineering computation problems are usually complex and consequently the development of parallel programs for solving these problems is a difficult task. In this paper, we describe the graph-oriented programming (GOP) model and environment for building and evaluating parallel applications. The GOP model provides higher level abstractions for message-passing parallel programming and the software environment offers tools which can ease programmers for parallelizing, writing, and deploying scientific and engineering computing applications. We discuss the motivations and various issues in developing the model and the software environment, present the design of the system architecture and the components, and describe the evaluation of the environment implemented on top of MPI with a sample parallel scientific application program. With the support of the high-level abstractions provided by the proposed GOP environment, programming of parallel applications on various parallel architectures can be greatly simplified.
π SIMILAR VOLUMES
We present a new computing approach for the parallelization on message-passing computer architectures of the DNAml algorithm, one of the most powerful tools available for constructing phylogenetic trees from DNA sequences. An analysis of the data dependencies of the method gave little chances to dev
## Abstract This paper proposes a trace reduction method for assessing the improvability of the performance of message passing parallel programs. This assessment is based on a whatβif prediction approach that forecasts future program performance, for example, the execution time if the target progra
Today's massively parallel machines are typically message-passing systems consisting of hundreds or thousands of processors. Implementing parallel applications efficiently in this environment is a challenging task, and poor parallel design decisions can be expensive to correct. Tools and techniques