Influence of weak labels for emotion recognition of tweets

5Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Research on emotion recognition of tweets focuses on feature engineering or algorithm design, while dataset labels are barely questioned. Datasets of tweets are often labelled manually or via crowdsourcing, which results in strong labels. These methods are time intensive and can be expensive. Alternatively, tweet hashtags can be used as free, inexpensive weak labels. This paper investigates the impact of using weak labels compared to strong labels. The study uses two label sets for a corpus of tweets. The weakly annotated label set is created employing the hashtags of the tweets, while the strong label set is created by the use of crowdsourcing. Both label sets are used separately as input for five classification algorithms to determine the classification performance of the weak labels. The results indicate only a 9.25% decrease in f1-score when using weak labels. This performance decrease does not outweigh the benefits of having free labels.

Cite

CITATION STYLE

APA

Janssens, O., Verstockt, S., Mannens, E., Van Hoecke, S., & Van De Walle, R. (2014). Influence of weak labels for emotion recognition of tweets. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8891, pp. 108–118). Springer Verlag. https://doi.org/10.1007/978-3-319-13817-6_12

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free