Identifying beneficial task relations for multi-task learning in deep neural networks

147Citations
Citations of this article
267Readers
Mendeley users who have this article in their library.

Abstract

Multi-task learning (MTL) in deep neural networks for NLP has recently received increasing interest due to some compelling benefits, including its potential to efficiently regularize models and to reduce the need for labeled data. While it has brought significant improvements in a number of NLP tasks, mixed results have been reported, and little is known about the conditions under which MTL leads to gains in NLP. This paper sheds light on the specific task relations that can lead to gains from MTL models over single-task setups.

Cite

CITATION STYLE

APA

Binge, J., & Sogaard, A. (2017). Identifying beneficial task relations for multi-task learning in deep neural networks. In 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017 - Proceedings of Conference (Vol. 2, pp. 164–169). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/e17-2026

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free