GPU-accelerated language and communication support by FPGA

0Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Although the GPU is one of the most successfully used accelerating devices for HPC, there are several issues when it is used for large-scale parallel systems. To describe real applications on GPU-ready parallel systems, we need to combine different paradigms of programming such as CUDA/OpenCL, MPI, and OpenMP for advanced platforms. In the hardware configuration, inter-GPU communication through PCIe channel and support by CPU are required which causes large overhead to be a bottleneck of total parallel processing performance. In our project to be described in this chapter, we developed an FPGA-based platform to reduce the latency of inter-GPU communication and also a PGAS language for distributed-memory programming with accelerating devices such as GPU. Through this work, a new approach to compensate the hardware and software weakness of parallel GPU computing is provided. Moreover, FPGA technology for computation and communication acceleration is described upon astrophysical problem where GPU or CPU computation is not sufficient on performance.

Cite

CITATION STYLE

APA

Boku, T., Hanawa, T., Murai, H., Nakao, M., Miki, Y., Amano, H., & Umemura, M. (2018). GPU-accelerated language and communication support by FPGA. In Advanced Software Technologies for Post-Peta Scale Computing: The Japanese Post-Peta CREST Research Project (pp. 301–317). Springer Singapore. https://doi.org/10.1007/978-981-13-1924-2_15

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free