Document classification by inversion of distributed language representations

27Citations
Citations of this article
202Readers
Mendeley users who have this article in their library.
Get full text

Abstract

There have been many recent advances in the structure and measurement of dis-tributed language models: those that map from words to a vector-space that is rich in information about word choice and com-position. This vector-space is the dis-tributed language representation. The goal of this note is to point out that any distributed representation can be turned into a classifier through inversion via Bayes rule. The approach is simple and modular, in that it will work with any language representation whose train-ing can be formulated as optimizing a probability model. In our application to 2 million sentences from Yelp reviews, we also find that it performs as well as or bet-ter than complex purpose-built algorithms.

Cite

CITATION STYLE

APA

Taddy, M. (2015). Document classification by inversion of distributed language representations. In ACL-IJCNLP 2015 - 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing, Proceedings of the Conference (Vol. 2, pp. 45–49). Association for Computational Linguistics (ACL). https://doi.org/10.3115/v1/p15-2008

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free