*-CFQ: Analyzing the Scalability of Machine Learning on a Compositional Task

10Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.

Abstract

We present *-CFQ (“star-CFQ”): a suite of large-scale datasets of varying scope based on the CFQ semantic parsing benchmark, designed for principled investigation of the scalability of machine learning systems in a realistic compositional task setting. Using this suite, we conduct a series of experiments investigating the ability of Transformers to benefit from increased training size under conditions of fixed computational cost. We show that compositional generalization remains a challenge at all training sizes, and we show that increasing the scope of natural language leads to consistently higher error rates, which are only partially offset by increased training data. We further show that while additional training data from a related domain improves the accuracy in data-starved situations, this improvement is limited and diminishes as the distance from the related domain to the target domain increases.

Cite

CITATION STYLE

APA

Tsarkov, D., Tihon, T., Scales, N., Momchev, N., Sinopalnikov, D., & Schärli, N. (2021). *-CFQ: Analyzing the Scalability of Machine Learning on a Compositional Task. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 11B, pp. 9949–9957). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i11.17195

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free