Umlaut: Debugging deep learning programs using program structure and model behavior

44Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Training deep neural networks can generate non-descriptive error messages or produce unusual output without any explicit errors at all. While experts rely on tacit knowledge to apply debugging strategies, non-experts lack the experience required to interpret model output and correct Deep Learning (DL) programs. In this work, we identify DL debugging heuristics and strategies used by experts, and use them to guide the design of Umlaut. Umlaut checks DL program structure and model behavior against these heuristics; provides human-readable error messages to users; and annotates erroneous model output to facilitate error correction. Umlaut links code, model output, and tutorial-driven error messages in a single interface. We evaluated Umlaut in a study with 15 participants to determine its efectiveness in helping developers fnd and fx errors in their DL programs. Participants using Umlaut found and fxed signifcantly more bugs compared to a baseline condition.

Cite

CITATION STYLE

APA

Schoop, E., Huang, F., & Hartmann, B. (2021). Umlaut: Debugging deep learning programs using program structure and model behavior. In Conference on Human Factors in Computing Systems - Proceedings. Association for Computing Machinery. https://doi.org/10.1145/3411764.3445538

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free