Managing Randomness to Enable Reproducible Machine Learning

8Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The National Information Standards Organization defines scientific reproducibility as "obtaining consistent results using the same input data, computational steps, methods, and code, and conditions of analysis'' [12] reproducibility. Reproducibility in machine learning (ML) refers to the ability to regenerate an ML model precisely guaranteeing identical accuracy and transparency. While a model may offer reproducible inference, reproducing the model itself is frequently problematic at best due to the presence of pseudo-random numbers as part of the model generation. One way to ensure that models are trustworthy is by managing the random numbers produced during model training. This paper establishes examples of the impact of randomness in model generation and offers a preliminary investigation into how random number generation can be controlled to make ML models more reproducible.

Cite

CITATION STYLE

APA

Ahmed, H., & Lofstead, J. (2022). Managing Randomness to Enable Reproducible Machine Learning. In P-RECS 2022 - Proceedings of the 5th International Workshop on Practical Reproducible Evaluation of Computer Systems, co-located with HPDC 2022 (pp. 15–20). Association for Computing Machinery, Inc. https://doi.org/10.1145/3526062.3536353

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free