Abstract
The process of obtaining high quality labeled data for natural language understanding tasks is often slow, error-prone, complicated and expensive. With the vast usage of neural networks, this issue becomes more notorious since these networks require a large amount of labeled data to produce satisfactory results. We propose a methodology to blend high quality but scarce labeled data with noisy but abundant weak labeled data during the training of neural networks. Experiments in the context of topic-dependent evidence detection with two forms of weak labeled data show the advantages of the blending scheme. In addition, we provide a manually annotated data set for the task of topic-dependent evidence detection.
Cite
CITATION STYLE
Shnarch, E., Alzate, C., Dankin, L., Gleize, M., Hou, Y., Choshen, L., … Slonim, N. (2018). Will it blend? Blending weak and strong labeled data in a neural network for argumentation mining. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 2, pp. 599–605). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-2095
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.