Batch matrix exponentiation

3Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Matrix-matrix multiplication can be considered a linchpin of applied numerical dense linear algebra as the performance of many common dense linear algebra packages is closely tied to the performance of matrix-matrix multiplication. Batch matrix-matrix multiplication, the matrix-matrix multiplication of a large number of relatively small matrices, is a developing area within dense linear algebra and is relevant to various application areas such as phylogenetics, finite element modeling, image processing, fluid dynamics, and hydrodynamics. Using batch matrix-matrix multiplication as the foundation, we have developed an optimized batch matrix exponentiation algorithm in CUDA that outperforms cublasXgemmBatched for small square matrices. After introducing the original motivation for our problem, matrix exponentiation from the phylogenetics domain, we discuss our algorithm in the context of both cublasXgemmBatched, and two alternative GPU methods for the numerical computation of matrix exponentiation: Lagrange interpolation, and Newton interpolation. All comparisons are done on both the Fermi and the Kepler architectures.

Cite

CITATION STYLE

APA

Lopez, M. G., & Horton, M. D. (2014). Batch matrix exponentiation. In Numerical Computations with GPUs (pp. 45–67). Springer International Publishing. https://doi.org/10.1007/978-3-319-06548-9_3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free