Abstract
We present Plato, a probabilistic model for entity resolution that includes a novel approach for handling noisy or uninformative features, and supplements labeled training data derived from Wikipedia with a very large unlabeled text corpus. Training and inference in the proposed model can easily be distributed across many servers, allowing it to scale to over 10 7 entities. We evaluate Plato on three standard datasets for entity resolution. Our approach achieves the best results to-date on TAC KBP 2011 and is highly competitive on both the CoNLL 2003 and TAC KBP 2012 datasets.
Cite
CITATION STYLE
Lazic, N., Subramanya, A., Ringgaard, M., & Pereira, F. (2015). Plato: A Selective Context Model for Entity Resolution. Transactions of the Association for Computational Linguistics, 3, 503–515. https://doi.org/10.1162/tacl_a_00154
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.