Prior literature on adversarial attack methods has mainly focused on attacking with and defending against a single threat model, e.g., perturbations bounded in Lp ball. However, multiple threat models can be combined into composite perturbations. One such approach, composite adversarial attack (CAA), not only expands the perturbable space of the image, but also may be overlooked by current modes of robustness evaluation. This paper demonstrates how CAA's attack order affects the resulting image, and provides real-time inferences of different models, which will facilitate users' configuration of the parameters of the attack level and their rapid evaluation of model prediction. A leaderboard to benchmark adversarial robustness against CAA is also introduced.
CITATION STYLE
Hsiung, L., Tsai, Y. Y., Chen, P. Y., & Ho, T. Y. (2022). CARBEN: Composite Adversarial Robustness Benchmark. In IJCAI International Joint Conference on Artificial Intelligence (pp. 5908–5911). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2022/851
Mendeley helps you to discover research relevant for your work.