Adversarial Reprogramming of Pretrained Neural Networks for Fraud Detection

12Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Machine learning models have been widely used for fraud detection, while developing and maintaining these models often suffers from significant limitations in terms of training data scarcity and constrained resources. To address these issues, in this paper, we leverage machine learning vulnerability to adversarial attacks, and design a novel model AdvRFD that Adversarially Reprograms an ImageNet classification neural network for Fraud Detection task. AdvRFD first embeds transaction features into a host image to construct new ImageNet data, and then learns a universal perturbation to be added to all inputs, such that the outputs of the pretrained model can be accordingly mapped to the final detection decisions for all transactions. Extensive experiments on two transaction datasets made over Ethereum and credit cards have demonstrated that AdvRFD is effective to detect fraud using limited data and resources.

Cite

CITATION STYLE

APA

Chen, L., Fan, Y., & Ye, Y. (2021). Adversarial Reprogramming of Pretrained Neural Networks for Fraud Detection. In International Conference on Information and Knowledge Management, Proceedings (pp. 2935–2939). Association for Computing Machinery. https://doi.org/10.1145/3459637.3482053

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free