Bootstrapping generators from noisy data

29Citations
Citations of this article
129Readers
Mendeley users who have this article in their library.

Abstract

A core step in statistical data-To-Text generation concerns learning correspondences between structured data representations (e.g., facts in a database) and associated texts. In this paper we aim to bootstrap generators from large scale datasets where the data (e.g., DBPedia facts) and related texts (e.g., Wikipedia abstracts) are loosely aligned. We tackle this challenging task by introducing a special-purpose content selection mechanism.1 We use multi-instance learning to automatically discover correspondences between data and text pairs and show how these can be used to enhance the content signal while training an encoder-decoder architecture. Experimental results demonstrate that models trained with content-specific objectives improve upon a vanilla encoder-decoder which solely relies on soft attention.

Cite

CITATION STYLE

APA

Perez-Beltrachini, L., & Lapata, M. (2018). Bootstrapping generators from noisy data. In NAACL HLT 2018 - 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference (Vol. 1, pp. 1516–1527). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/n18-1137

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free