Learning Universal Authorship Representations

32Citations
Citations of this article
61Readers
Mendeley users who have this article in their library.

Abstract

Determining whether two documents were composed by the same author, also known as authorship verification, has traditionally been tackled using statistical methods. Recently, authorship representations learned using neural networks have been found to outperform alternatives, particularly in large-scale settings involving hundreds of thousands of authors. But do such representations learned in a particular domain transfer to other domains? Or are these representations inherently entangled with domain-specific features? To study these questions, we conduct the first large-scale study of cross-domain transfer for authorship verification considering zero-shot transfers involving three disparate domains: Amazon reviews, fanfiction short stories, and Reddit comments. We find that although a surprising degree of transfer is possible between certain domains, it is not so successful between others. We examine properties of these domains that influence generalization and propose simple but effective methods to improve transfer.

Cite

CITATION STYLE

APA

Rivera Soto, R. A., Miano, O., Ordonez, J., Chen, B., Khan, A., Bishop, M., & Andrews, N. (2021). Learning Universal Authorship Representations. In EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 913–919). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.emnlp-main.70

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free