The National Information Standards Organization defines scientific reproducibility as "obtaining consistent results using the same input data, computational steps, methods, and code, and conditions of analysis'' [12] reproducibility. Reproducibility in machine learning (ML) refers to the ability to regenerate an ML model precisely guaranteeing identical accuracy and transparency. While a model may offer reproducible inference, reproducing the model itself is frequently problematic at best due to the presence of pseudo-random numbers as part of the model generation. One way to ensure that models are trustworthy is by managing the random numbers produced during model training. This paper establishes examples of the impact of randomness in model generation and offers a preliminary investigation into how random number generation can be controlled to make ML models more reproducible.
CITATION STYLE
Ahmed, H., & Lofstead, J. (2022). Managing Randomness to Enable Reproducible Machine Learning. In P-RECS 2022 - Proceedings of the 5th International Workshop on Practical Reproducible Evaluation of Computer Systems, co-located with HPDC 2022 (pp. 15–20). Association for Computing Machinery, Inc. https://doi.org/10.1145/3526062.3536353
Mendeley helps you to discover research relevant for your work.