Task-specific representation learning for web-scale entity disambiguation

6Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

Named entity disambiguation (NED) is a central problem in information extraction. The goal is to link entities in a knowledge graph (KG) to their mention spans in unstructured text. Each distinct mention span (like John Smith, Jordan or Apache) represents a multi-class classification task. NED can therefore be modeled as a multitask problem with tens of millions of tasks for realistic KGs. We initiate an investigation into neural representations, network architectures, and training protocols for multitask NED. Specifically, we propose a task-sensitive representation learning framework that learns mention dependent representations, followed by a common classifier. Parameter learning in our framework can be decomposed into solving multiple smaller problems involving overlapping groups of tasks. We prove bounds for excess risk, which provide additional insight into the problem of multi-task representation learning. While remaining practical in terms of training memory and time requirements, our approach outperforms recent strong baselines, on four benchmark data sets.

Cite

CITATION STYLE

APA

Kar, R., Reddy, S., Bhattacharya, S., Dasgupta, A., & Chakrabarti, S. (2018). Task-specific representation learning for web-scale entity disambiguation. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 5812–5819). AAAI press. https://doi.org/10.1609/aaai.v32i1.12066

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free