Explainability-based Debugging of Machine Learning for Vulnerability Discovery

14Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

Machine learning has been successfully used for increasingly complex and critical tasks, achieving high performance and efficiency that would not be possible for human operators. Unfortunately, recent studies have shown that, despite its power, this technology tends to learn spurious correlations from data, making it weak and susceptible to manipulation. Explainability techniques are often used to identify the most relevant features contributing to the decision. However, this is often done by taking examples one by one and trying to show the problem locally. To mitigate this issue, we propose in this paper a systematic method to leverage explainability techniques and build on their results to highlight problems in the model design and training. With an empirical analysis on the Devign dataset, we validate the proposed methodology with a CodeBERT model trained for vulnerability discovery, showing that, despite its impressive performances, spurious correlations consistently steer its decision.

Cite

CITATION STYLE

APA

Sotgiu, A., Pintor, M., & Biggio, B. (2022). Explainability-based Debugging of Machine Learning for Vulnerability Discovery. In ACM International Conference Proceeding Series. Association for Computing Machinery. https://doi.org/10.1145/3538969.3543809

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free