Abstract interpretation of decision tree ensemble classifiers

30Citations
Citations of this article
32Readers
Mendeley users who have this article in their library.

Abstract

We study the problem of formally and automatically verifying robustness properties of decision tree ensemble classifiers such as random forests and gradient boosted decision tree models. A recent stream of works showed how abstract interpretation, which is ubiquitously used in static program analysis, can be successfully deployed to formally verify (deep) neural networks. In this work we push forward this line of research by designing a general and principled abstract interpretation-based framework for the formal verification of robustness and stability properties of decision tree ensemble models. Our abstract interpretation-based method may induce complete robustness checks of standard adversarial perturbations and output concrete adversarial attacks. We implemented our abstract verification technique in a tool called silva, which leverages an abstract domain of not necessarily closed real hyperrectangles and is instantiated to verify random forests and gradient boosted decision trees. Our experimental evaluation on the MNIST dataset shows that silva provides a precise and efficient tool which advances the current state of the art in tree ensembles verification.

Cite

CITATION STYLE

APA

Ranzato, F., & Zanella, M. (2020). Abstract interpretation of decision tree ensemble classifiers. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 5478–5486). AAAI press. https://doi.org/10.1609/aaai.v34i04.5998

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free