Is It Smaller Than a Tennis Ball? Language Models Play the Game of Twenty Questions

3Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Researchers often use games to analyze the abilities of Artificial Intelligence models. In this work, we use the game of Twenty Questions to study the world knowledge of language models. Despite its simplicity for humans, this game requires a broad knowledge of the world to answer yes/no questions. We evaluate several language models on this task and find that only the largest model has enough world knowledge to play it well, although it still has difficulties with the shape and size of objects. We also present a new method to improve the knowledge of smaller models by leveraging external information from the web. Finally, we release our dataset and Twentle, a website to interactively test the knowledge of language models by playing Twenty Questions.

Cite

CITATION STYLE

APA

De Bruyn, M., Lotfi, E., Buhmann, J., & Daelemans, W. (2022). Is It Smaller Than a Tennis Ball? Language Models Play the Game of Twenty Questions. In BlackboxNLP 2022 - BlackboxNLP Analyzing and Interpreting Neural Networks for NLP, Proceedings of the Workshop (pp. 80–90). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.blackboxnlp-1.7

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free