Adversarial attack on machine learning models

0Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Machine Learning (ML) models are applied in a variety of tasks such as network intrusion detection or malware classification. Yet, these models are vulnerable to a class of malicious inputs known as adversarial examples. These are slightly perturbed inputs that are classified incorrectly by the ML model. The mitigation of these adversarial inputs remains an open problem. As a step towards understanding adversarial examples, we show that they are not drawn from the same distribution than the original data, and can thus be detected using statistical tests. Using this knowledge, we introduce a complimentary approach to identify specific inputs that are adversarial. Specifically, we augment our ML model with an additional output, in which the model is trained to classify all adversarial inputs.

Cite

CITATION STYLE

APA

Sahaya Sakila, V., Sandeep, M., & Praveen Hari Krishna, N. (2019). Adversarial attack on machine learning models. International Journal of Innovative Technology and Exploring Engineering, 8(6 Special Issue 4), 431–434. https://doi.org/10.35940/ijitee.F1088.0486S419

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free