Abstract
Graphic processing Units (GPUs) are gaining ground in high-performance computing. CUDA (an extension to C) is most widely used parallel programming framework for general purpose GPU computations. However, the task of writing optimized CUDA program is complex even for experts. We present a method for restructuring loops into an optimized CUDA kernels based on a 3-step algorithm which are loop tiling, coalesced memory access, and resource optimization. We also establish the relationships between the influencing parameters and propose a method for finding possible tiling solutions with coalesced memory access that best meets the identified constraints. We also present a simplified algorithm for restructuring loops and rewrite them as an efficient CUDA Kernel. The execution model of synthesized kernel consists of uniformly distributing the kernel threads to keep all cores busy while transferring a tailored data locality which is accessed using coalesced pattern to amortize the long latency of the secondary memory. In the evaluation, we implement some simple applications using the proposed restructuring strategy and evaluate the performance in terms of execution time and GPU throughput. © 2012 IEEE.
Original language | English (US) |
---|---|
Title of host publication | 2012 2nd IEEE International Conference on Parallel, Distributed and Grid Computing |
Publisher | Institute of Electrical and Electronics Engineers (IEEE) |
Pages | 55-60 |
Number of pages | 6 |
ISBN (Print) | 9781467329255 |
DOIs | |
State | Published - Dec 2012 |
Externally published | Yes |
Bibliographical note
KAUST Repository Item: Exported on 2020-10-01Acknowledgements: Thanks to the ICS-KFUPM and KAUST for givingaccess to their GPU computers and workstations.
This publication acknowledges KAUST support, but has no KAUST affiliated authors.