Automatic CUDA code synthesis framework for multicore CPU and GPU architectures

4Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recently, general purpose GPU (GPGPU) programming has spread rapidly after CUDA was first introduced to write parallel programs in high-level languages for NVIDIA GPUs. While a GPU exploits data parallelism very effectively, task-level parallelism is exploited as a multi-threaded program on a multicore CPU. For such a heterogeneous platform that consists of a multicore CPU and GPU, we propose an automatic code synthesis framework that takes a process network model specification as input and generates a multithreaded CUDA code. With the model based specification, one can explicitly specify both function-level and loop-level parallelism in an application and explore the wide design space in mapping of function blocks and selecting the communication methods between CPU and GPU. The proposed technique is complementary to other high-level methods of CUDA programming. © 2012 Springer-Verlag.

Cite

CITATION STYLE

APA

Jung, H., Yi, Y., & Ha, S. (2012). Automatic CUDA code synthesis framework for multicore CPU and GPU architectures. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7203 LNCS, pp. 579–588). https://doi.org/10.1007/978-3-642-31464-3_59

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free