Abstract
Machine learning has been successfully used for increasingly complex and critical tasks, achieving high performance and efficiency that would not be possible for human operators. Unfortunately, recent studies have shown that, despite its power, this technology tends to learn spurious correlations from data, making it weak and susceptible to manipulation. Explainability techniques are often used to identify the most relevant features contributing to the decision. However, this is often done by taking examples one by one and trying to show the problem locally. To mitigate this issue, we propose in this paper a systematic method to leverage explainability techniques and build on their results to highlight problems in the model design and training. With an empirical analysis on the Devign dataset, we validate the proposed methodology with a CodeBERT model trained for vulnerability discovery, showing that, despite its impressive performances, spurious correlations consistently steer its decision.
Author supplied keywords
Cite
CITATION STYLE
Sotgiu, A., Pintor, M., & Biggio, B. (2022). Explainability-based Debugging of Machine Learning for Vulnerability Discovery. In ACM International Conference Proceeding Series. Association for Computing Machinery. https://doi.org/10.1145/3538969.3543809
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.