Syntactic language models and N-gram language models have both been used in word ordering. In this paper, we give an empirical comparison between N-gram and syntactic language models on word order task. Our results show that the quality of automatically-parsed training data has a relatively small impact on syntactic models. Both of syntactic and N-gram models can benefit from large-scale raw text. Compared with N-gram models, syntactic models give overall better performance, but they require much more training time. In addition, the two models lead to different error distributions in word ordering. A combination of the two models integrates the advantages of each model, achieving the best result in a standard benchmark.
CITATION STYLE
Liu, J., & Zhang, Y. (2015). An empirical comparison between N-gram and syntactic language models for word ordering. In Conference Proceedings - EMNLP 2015: Conference on Empirical Methods in Natural Language Processing (pp. 369–378). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d15-1043
Mendeley helps you to discover research relevant for your work.