Effects of Adversarial Training on the Safety of Classification Models

1Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

Artificial intelligence (AI) is one of the most important topics that implements symmetry in computer science. As like humans, most AI also learns by trial-and-error approach which requires appropriate adversarial examples. In this study, we prove that adversarial training can be useful to verify the safety of classification model in early stage of development. We experimented with various amount of adversarial data and found that the safety can be significantly improved by appropriate ratio of adversarial training.

Cite

CITATION STYLE

APA

Kim, H., & Han, J. (2022). Effects of Adversarial Training on the Safety of Classification Models. Symmetry, 14(7). https://doi.org/10.3390/sym14071338

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free