A parallel distributed implementation of the second-order Mdler-Plesset perturbation theory method, widely used in quantum chemistry, is presented. Parallelization strategy and performance for the HONDO quantum chemistry program running on a network of Unix computers are also discussed. Superlinear
Event parallelism: Distributed memory parallel computing for high energy physics experiments
β Scribed by Thomas Nash
- Publisher
- Elsevier Science
- Year
- 1989
- Tongue
- English
- Weight
- 927 KB
- Volume
- 57
- Category
- Article
- ISSN
- 0010-4655
No coin nor oath required. For personal study only.
β¦ Synopsis
This paper describes the present and expected future development of distributed memory parallel computers for high energy physics experiments. It covers the use of event parallel microprocessor farms, particularly at Fermilab, including both ACP multiprocessors and farms of MicroVAXES. These systems have proven very cost effective in the past. A case is made for moving to the more open environment of UNIX and RISC processors. The 2nd Generation ACP Multiprocessor System, which is based on powerful RISC systems, is described. Given the promise of still more extraordinary increases in processor performance, a new emphasis on point to point, rather than bussed, communication will be required. Developments in this direction are described.
π SIMILAR VOLUMES
Ray tracing is a well known technique to generate life-like images. Unfortunately, ray tracing complex scenes can require large amounts of CPU time and memory storage. Distributed memory parallel computers with large memory capacities and high processing speeds are ideal candidates to perform ray tr
A parallel implementation of the computation of RHF energy second derivatives with respect to the nuclear coordinates is described. The algorithm and organization of the code are described in detail on the most computationally demanding steps with special emphasis on the integral transformation code
This paper introduces an architecture-independent, hierarchical approach to algorithm design on distributed-memory architectures, in contrast to the current trend of tailoring algorithms towards specific architectures. We show that, rather surprisingly, this new approach can achieve uniformity witho
To parallelize applications that require the use of random numbers, an efficient and good quality parallel random number generator is required. In this paper, we study the parallelization of lagged Fibonacci generators for distributed memory parallel computers. Two popular ways of generating a rando
Parallel cluster architectures are an effective way to build high-performance computer systems at a moderate price. Consequently, clusters will become very attractive to parallel applications and high-end parallel servers, which are usually built around distributed object middleware. Commonly used m