Scaling address parsing sequence models through active learning

8Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Address parsing is a critical step for map search engines. This component annotates the terms of an address query, e.g. house numbers, road names, administrative units etc., so that the address search engine can resolve the expected result. Deep recurrent models achieve state of the art performance for address parsing; however, scaling such models is problematic. They require a significant amount of term-annotated data which is expensive to acquire. In this paper, active learning significantly reduces the amount of labeled data required to train accurate address parsing models. We demonstrate the efficiency of our approach when cold-starting with human-labeled as well as synthetically-generated data.

Cite

CITATION STYLE

APA

Craig, H., Yankov, D., Wang, R., Berkhin, P., & Wu, W. (2019). Scaling address parsing sequence models through active learning. In GIS: Proceedings of the ACM International Symposium on Advances in Geographic Information Systems (pp. 424–427). Association for Computing Machinery. https://doi.org/10.1145/3347146.3359070

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free