Abstract
Training deep neural networks can generate non-descriptive error messages or produce unusual output without any explicit errors at all. While experts rely on tacit knowledge to apply debugging strategies, non-experts lack the experience required to interpret model output and correct Deep Learning (DL) programs. In this work, we identify DL debugging heuristics and strategies used by experts, and use them to guide the design of Umlaut. Umlaut checks DL program structure and model behavior against these heuristics; provides human-readable error messages to users; and annotates erroneous model output to facilitate error correction. Umlaut links code, model output, and tutorial-driven error messages in a single interface. We evaluated Umlaut in a study with 15 participants to determine its efectiveness in helping developers fnd and fx errors in their DL programs. Participants using Umlaut found and fxed signifcantly more bugs compared to a baseline condition.
Author supplied keywords
Cite
CITATION STYLE
Schoop, E., Huang, F., & Hartmann, B. (2021). Umlaut: Debugging deep learning programs using program structure and model behavior. In Conference on Human Factors in Computing Systems - Proceedings. Association for Computing Machinery. https://doi.org/10.1145/3411764.3445538
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.