Title of program. PRACAH in CPC. This program is different in that it is a parallel solution programmed in OCCAM and executed on a network Catalogue number: AAXE of TRANSPUTERS. Program obtainable from: CPC Program Library, Queen's Urn-Restrictions on the complexity of the problem versity of Belfast
Effective data parallel computation using the Psi calculus
โ Scribed by Mullin, L.M.R.; Jenkins, M.A.
- Book ID
- 102645559
- Publisher
- John Wiley and Sons
- Year
- 1996
- Tongue
- English
- Weight
- 861 KB
- Volume
- 8
- Category
- Article
- ISSN
- 1040-3108
No coin nor oath required. For personal study only.
โฆ Synopsis
Large scale scientific computing necessitates finding a way to match the high level understanding of how a problem can be solved with the details of its computation in a processing environment organized as networks of processors. Effective utilization of parallel architectures can then be achieved by using formal methods to describe both computations and computational organizations within these networks. By returning to the mathematical treatment of a problem as a high level numerical algorithm we can express it as an algorithmic formalism that captures the inherent parallelism of the computation. We then give a meta description of an architecture followed by the use of transformational techniques to convert the high level description into a program that utilizes the architecture effectively. The hope is that one formalism can be used to describe both computations as well as architectures and that a methodology for automatically transforming computations can be developed. The formalism and methodology presented in the paper is a first step toward the ambitious goals described above. It uses a theory of arrays, the Psi calculus, as the formalism, and two levels of conversionsone for simplification and another for data mapping.
๐ SIMILAR VOLUMES
However, the speedup achieved through parallelism is often lower in modern systems. It is no surprise, then, that developers of compilers for data-parallel languages have hypothesized the importance of optimizations that overlap communications with computations in order to reduce execution times and
This paper addresses optimal mapping of parallel programs composed of a chain of data parallel tasks onto the processors of a parallel system. The input to the programs is a stream of data sets, each of which is processed in order by the chain of tasks. This computation structure, also referred to a
Arrays are mapped to processors through a two-step process-alignment followed by distribution-in data-parallel languages such as High Performance Fortran. This process of mapping creates disjoint pieces of the array that are locally owned by each processor. An HPF compiler that generates code for ar