Abstract
Large-scale semantic parsing datasets annotated with logical forms have enabled major advances in supervised approaches. But can richer supervision help even more? To explore the utility of fine-grained, lexical-level supervision, we introduce SQUALL, a dataset that enriches 11,276 WIKITABLEQUESTIONS English-language questions with manually created SQL equivalents plus alignments between SQL and question fragments. Our annotation enables new training possibilities for encoder-decoder models, including approaches from machine translation previously precluded by the absence of alignments. We propose and test two methods: (1) supervised attention; (2) adopting an auxiliary objective of disambiguating references in the input queries to table columns. In 5-fold cross validation, these strategies improve over strong baselines by 4.4% execution accuracy. Oracle experiments suggest that annotated alignments can support further accuracy gains of up to 23.9%.
Cite
CITATION STYLE
Shi, T., Zhao, C., Boyd-Graber, J., Daumé, H., & Lee, L. (2020). On the potential of lexico-logical alignments for semantic parsing to SQL queries. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 1849–1864). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.167
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.