A survey of application memory usage on a national supercomputer: An analysis of memory requirements on ARCHER

4Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this short paper we set out to provide a set of modern data on the actual memory per core and memory per node requirements of the most heavily used applications on a contemporary, national-scale supercomputer. This report is based on data from all jobs run on the UK national supercomputing service, ARCHER, a 118,000 core Cray XC30, in the 1 year period from 1st July 2016 to 30th June 2017 inclusive. Our analysis shows that 80% of all usage on ARCHER has a maximum memory use of 1 GiB/core or less (24 GiB/node or less) and that there is a trend to larger memory use as job size increases. Analysis of memory use by software application type reveals differences in memory use between periodic electronic structure, atomistic N-body, grid-based climate modelling, and grid-based CFD applications. We present an analysis of these differences, and suggest further analysis and work in this area. Finally, we discuss the implications of these results for the design of future HPC systems, in particular the applicability of high bandwidth memory type technologies.

Author supplied keywords

Cite

CITATION STYLE

APA

Turner, A., & McIntosh-Smith, S. (2018). A survey of application memory usage on a national supercomputer: An analysis of memory requirements on ARCHER. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10724 LNCS, pp. 250–260). Springer Verlag. https://doi.org/10.1007/978-3-319-72971-8_13

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free