Named entity disambiguation (NED) is a central problem in information extraction. The goal is to link entities in a knowledge graph (KG) to their mention spans in unstructured text. Each distinct mention span (like John Smith, Jordan or Apache) represents a multi-class classification task. NED can therefore be modeled as a multitask problem with tens of millions of tasks for realistic KGs. We initiate an investigation into neural representations, network architectures, and training protocols for multitask NED. Specifically, we propose a task-sensitive representation learning framework that learns mention dependent representations, followed by a common classifier. Parameter learning in our framework can be decomposed into solving multiple smaller problems involving overlapping groups of tasks. We prove bounds for excess risk, which provide additional insight into the problem of multi-task representation learning. While remaining practical in terms of training memory and time requirements, our approach outperforms recent strong baselines, on four benchmark data sets.
CITATION STYLE
Kar, R., Reddy, S., Bhattacharya, S., Dasgupta, A., & Chakrabarti, S. (2018). Task-specific representation learning for web-scale entity disambiguation. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 5812–5819). AAAI press. https://doi.org/10.1609/aaai.v32i1.12066
Mendeley helps you to discover research relevant for your work.