Exploration of automatic optimization for CUDA programming

Type
Conference Paper

Authors
Al-Mouhamed, Mayez
Khan, Ayaz ul Hassan

Date
2012-12

Abstract
Graphic processing Units (GPUs) are gaining ground in high-performance computing. CUDA (an extension to C) is most widely used parallel programming framework for general purpose GPU computations. However, the task of writing optimized CUDA program is complex even for experts. We present a method for restructuring loops into an optimized CUDA kernels based on a 3-step algorithm which are loop tiling, coalesced memory access, and resource optimization. We also establish the relationships between the influencing parameters and propose a method for finding possible tiling solutions with coalesced memory access that best meets the identified constraints. We also present a simplified algorithm for restructuring loops and rewrite them as an efficient CUDA Kernel. The execution model of synthesized kernel consists of uniformly distributing the kernel threads to keep all cores busy while transferring a tailored data locality which is accessed using coalesced pattern to amortize the long latency of the secondary memory. In the evaluation, we implement some simple applications using the proposed restructuring strategy and evaluate the performance in terms of execution time and GPU throughput. © 2012 IEEE.

Citation
Al-Mouhamed M, Khan A ul H (2012) Exploration of automatic optimization for CUDA programming. 2012 2nd IEEE International Conference on Parallel, Distributed and Grid Computing. Available: http://dx.doi.org/10.1109/PDGC.2012.6449791.

Acknowledgements
Thanks to the ICS-KFUPM and KAUST for givingaccess to their GPU computers and workstations.

Publisher
Institute of Electrical and Electronics Engineers (IEEE)

Journal
2012 2nd IEEE International Conference on Parallel, Distributed and Grid Computing

DOI
10.1109/PDGC.2012.6449791

Permanent link to this record