The behavior and performance of MPI non-blocking message passing operations are sensitive to implementation specifics as they are heavily dependant on available system level buffers. In this paper we investigate the behavior of non-blocking communication primitives provided by popular MPI implementations and propose strategies for these primitives than can reduce processor synchronization overheads. We also demonstrate the improvements in the performance of a parallel Structured Adaptive Mesh Refinement (SAMR) application using these strategies. © Springer-Verlag 2004.
CITATION STYLE
Saif, T., & Parashar, M. (2004). Understanding the behavior and performance of non-blocking communications in MPI. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 3149, 173–182. https://doi.org/10.1007/978-3-540-27866-5_22
Mendeley helps you to discover research relevant for your work.