A Globally Normalized Neural Model for Semantic Parsing

2Citations
Citations of this article
45Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we propose a globally normalized model for context-free grammar (CFG)based semantic parsing. Instead of predicting a probability, our model predicts a real-valued score at each step and does not suffer from the label bias problem. Experiments show that our approach outperforms locally normalized models on small datasets, but it does not yield improvement on a large dataset.

Cite

CITATION STYLE

APA

Huang, C., Yang, W., Cao, Y., Zaïane, O., & Mou, L. (2021). A Globally Normalized Neural Model for Semantic Parsing. In SPNLP 2021 - 5th Workshop on Structured Prediction for NLP, Proceedings of the Workshop (pp. 61–66). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.spnlp-1.7

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free