TextHide: Tackling data privacy in language understanding tasks

35Citations
Citations of this article
100Readers
Mendeley users who have this article in their library.

Abstract

An unsolved challenge in distributed or federated learning is to effectively mitigate privacy risks without slowing down training or reducing accuracy. In this paper, we propose TextHide aiming at addressing this challenge for natural language understanding tasks. It requires all participants to add a simple encryption step to prevent an eavesdropping attacker from recovering private text data. Such an encryption step is efficient and only affects the task performance slightly. In addition, TextHide fits well with the popular framework of fine-tuning pre-trained language models (e.g., BERT) for any sentence or sentence-pair task. We evaluate TextHide on the GLUE benchmark, and our experiments show that TextHide can effectively defend attacks on shared gradients or representations and the averaged accuracy reduction is only 1.9%. We also present an analysis of the security of TextHide using a conjecture about the computational intractability of a mathematical problem.

Cite

CITATION STYLE

APA

Huang, Y., Song, Z., Chen, D., Li, K., & Arora, S. (2020). TextHide: Tackling data privacy in language understanding tasks. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 1368–1382). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.123

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free