FPGA vs. GPU for sparse matrix vector multiply

48Citations
Citations of this article
81Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Sparse matrix-vector multiplication (SpMV) is a common operation in numerical linear algebra and is the computational kernel of many scientific applications. It is one of the original and perhaps most studied targets for FPGA acceleration. Despite this, GPUs, which have only recently gained both general-purpose programmability and native support for double precision floating-point arithmetic, are viewed by some as a more effective platform for SpMV and similar linear algebra computations. In this paper, we present an analysis comparing an existing GPU SpMV implementation to our own, novel FPGA implementation. In this analysis, we describe the challenges faced by any SpMV implementation, the unique approaches to these challenges taken by both FPGA and GPU implementations, and their relative performance for SpMV. © 2009 IEEE.

Cite

CITATION STYLE

APA

Zhang, Y., Shalabi, Y. H., Jain, R., Nagar, K. K., & Bakos, J. D. (2009). FPGA vs. GPU for sparse matrix vector multiply. In Proceedings of the 2009 International Conference on Field-Programmable Technology, FPT’09 (pp. 255–262). https://doi.org/10.1109/FPT.2009.5377620

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free