An oral exam for measuring a dialog system's capabilities

3Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.

Abstract

This paper suggests a model and methodology for measuring the breadth and flexibility of a dialog system's capabilities. The approach relies on having human evaluators administer a targeted oral exam to a system and provide their subjective views of that system's performance on each test problem. We present results from one instantiation of this test being performed on two publicly-Accessible dialog systems and a human, and show that the suggested metrics do provide useful insights into the relative strengths and weaknesses of these systems. Results suggest that this approach can be performed with reasonable reliability and with reasonable amounts of effort. We hope that authors will augment their reporting with this approach to improve clarity and make more direct progress toward broadlycapable dialog systems.

Cite

CITATION STYLE

APA

Cohen, D., & Lane, I. (2016). An oral exam for measuring a dialog system’s capabilities. In 30th AAAI Conference on Artificial Intelligence, AAAI 2016 (pp. 835–841). AAAI press. https://doi.org/10.1609/aaai.v30i1.10060

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free