* New approaches to parallel computing are being developed that make better use of the heterogeneous cluster architecture * Provides a detailed introduction to parallel computing on heterogenous clusters * All concepts and algorithms are illustrated with working programs that can be compiled
Neural Network Parallel Computing
โ Scribed by Yoshiyasu Takefuji (auth.)
- Publisher
- Springer US
- Year
- 1992
- Tongue
- English
- Leaves
- 236
- Series
- The Springer International Series in Engineering and Computer Science 164
- Edition
- 1
- Category
- Library
No coin nor oath required. For personal study only.
โฆ Synopsis
Neural Network Parallel Computing is the first book available to the professional market on neural network computing for optimization problems. This introductory book is not only for the novice reader, but for experts in a variety of areas including parallel computing, neural network computing, computer science, communications, graph theory, computer aided design for VLSI circuits, molecular biology, management science, and operations research. The goal of the book is to facilitate an understanding as to the uses of neural network models in real-world applications.
Neural Network Parallel Computing presents a major breakthrough in science and a variety of engineering fields. The computational power of neural network computing is demonstrated by solving numerous problems such as N-queen, crossbar switch scheduling, four-coloring and k-colorability, graph planarization and channel routing, RNA secondary structure prediction, knight's tour, spare allocation, sorting and searching, and tiling.
Neural Network Parallel Computing is an excellent reference for researchers in all areas covered by the book. Furthermore, the text may be used in a senior or graduate level course on the topic.
โฆ Table of Contents
Front Matter....Pages i-xiii
Neural Network Models and N-Queen Problems....Pages 1-26
Crossbar Switch Scheduling Problems....Pages 27-36
Four-Coloring and K-Colorability Problems....Pages 37-50
Graph Planarization Problems....Pages 51-64
Channel Routing Problems....Pages 65-86
RNA Secondary Structure Prediction....Pages 87-109
Knightโs Tour Problems....Pages 111-118
Spare Allocation Problems....Pages 119-131
Sorting and Searching....Pages 133-144
Tiling Problems....Pages 145-156
Silicon Neural Networks....Pages 157-178
Mathematical Background....Pages 179-195
Forthcoming Applications....Pages 197-215
Conjunctoids and Artificial Learning....Pages 217-225
Back Matter....Pages 227-230
โฆ Subjects
Circuits and Systems;Statistical Physics, Dynamical Systems and Complexity;Electrical Engineering;Computer Science, general
๐ SIMILAR VOLUMES
<p>Understanding of the human brain functioning currently represents a challenging problem. In contrast to usual serial computers and complicated hierarchically organized artificial man-made systems, decentralized, parallel and distributed information processing principles are inherent to the brain.
Aims and Scope Understanding of the human brain functioning currently represents a challenging problem. In contrast to usual serial computers and complicated hierarchically organized artificial man-made systems, decentralized, parallel and distributed information processing principles are inheren
This highly acclaimed work, first published by Prentice Hall in 1989, is a comprehensive and theoretically sound treatment of parallel and distributed numerical methods. It focuses on algorithms that are naturally suited for massive parallelization, and it explores the fundamental convergence, rate
This highly acclaimed work, first published by Prentice Hall in 1989, is a comprehensive and theoretically sound treatment of parallel and distributed numerical methods. It focuses on algorithms that are naturally suited for massive parallelization, and it explores the fundamental convergence, rate
Neural Networks: Computational Models and Applications presents important theoretical and practical issues in neural networks, including the learning algorithms of feed-forward neural networks, various dynamical properties of recurrent neural networks, winner-take-all networks and their applications