BAYES-TREX: a Bayesian Sampling Approach to Model Transparency by Example

9Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.

Abstract

Post-hoc explanation methods are gaining popularity for interpreting, understanding, and debugging neural networks. Most analyses using such methods explain decisions in response to inputs drawn from the test set. However, the test set may have few examples that trigger some model behaviors, such as high-confidence failures or ambiguous classifications. To address these challenges, we introduce a flexible model inspection framework: BAYES-TREX. Given a data distribution, BAYES-TREX finds in-distribution examples which trigger a specified prediction confidence. We demonstrate several use cases of BAYES-TREX, including revealing highly confident (mis)classifications, visualizing class boundaries via ambiguous examples, understanding novel-class extrapolation behavior, and exposing neural network overconfidence. We use BAYES-TREX to study classifiers trained on CLEVR, MNIST, and Fashion-MNIST, and we show that this framework enables more flexible holistic model analysis than just inspecting the test set. Code and supplemental material are available at https://github.com/serenabooth/Bayes-TrEx.

Cite

CITATION STYLE

APA

Booth, S., Zhou, Y., Shah, A., & Shah, J. (2021). BAYES-TREX: a Bayesian Sampling Approach to Model Transparency by Example. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 13A, pp. 11423–11432). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i13.17361

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free