Parallelization of cellular neural networks on GPU
โ Scribed by Tze-Yui Ho; Ping-Man Lam; Chi-Sing Leung
- Publisher
- Elsevier Science
- Year
- 2008
- Tongue
- English
- Weight
- 983 KB
- Volume
- 41
- Category
- Article
- ISSN
- 0031-3203
No coin nor oath required. For personal study only.
โฆ Synopsis
Recently, cellular neural networks (CNNs) have been demonstrated to be a highly effective paradigm applicable in a wide range of areas. Typically, CNNs can be implemented using VLSI circuits, but this would unavoidably require additional hardware. On the other hand, we can also implement CNNs purely by software; this, however, would result in very low performance when given a large CNN problem size. Nowadays, conventional desktop computers are usually equipped with programmable graphics processing units (GPUs) that can support parallel data processing. This paper introduces a GPU-based CNN simulator. In detail, we carefully organize the CNN data as 4-channel textures, and efficiently implement the CNN computation as fragment programs running in parallel on a GPU. In this way, we can create a high performance but low-cost CNN simulator. Experimentally, we demonstrate that the resultant GPU-based CNN simulator can run 8-17 times faster than a CPU-based CNN simulator.
๐ SIMILAR VOLUMES
## ลฝ . A set of sufficient criteria are found ensuring the global asymptotic stability of delayed cellular neural networks DCNN with more general output functions by introducing ingeniously real parameters a ลฝ . s 1,a q b s 1,h q z s 1,h q z s 1 j s 1,2, PPP ,n , constructing suitable Lyapunov fu
In this paper is presented the use of the discrete-time cellular neural network (DTCNN) paradigm to develop algorithms devised for general-purpose massively parallel processing (MPP) systems. This paradigm is defined in discrete N-dimensional spaces (lattices) and is characterized by the locality of
The cellular neural network is able to perform different image-processing tasks depending on the template values, i.e. the network parameters, used. In the case of linear templates the parameter space is divided into different regions by hyperplanes. Every region is associated with a task, such that