We present a fast hybrid solver for dense linear systems based on LU factorization. To achieve good performance, we avoid pivoting by using random butterfly transformations for which we developed efficient implementations on heterogeneous architectures. We used both Graphics Processing Units and Intel Xeon Phi as accelerators. The performance results show that the pre-processing due to randomization is negligible and that the solver outperforms the corresponding routines based on partial pivoting.
CITATION STYLE
Baboulin, M., Khabou, A., & Rémy, A. (2015). A randomized Lu-based solver using GPU and Intel Xeon Phi accelerators. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9523, pp. 175–184). Springer Verlag. https://doi.org/10.1007/978-3-319-27308-2_15
Mendeley helps you to discover research relevant for your work.