Abstract
The availability of the Rhetorical Structure Theory (RST) Discourse Treebank has spurred substantial research into discourse analysis of written texts; however, limited research has been conducted to date on RST annotation and parsing of spoken language, in particular, nonnative spontaneous speech. Considering that the measurement of discourse coherence is typically a key metric in human scoring rubrics for assessments of spoken language, we initiated a research effort to obtain RST annotations of a large number of non-native spoken responses from a standardized assessment of academic English proficiency. The resulting inter-annotator ? agreements on the three different levels of Span, Nuclearity, and Relation are 0.848, 0.766, and 0.653, respectively. Furthermore, a set of features was explored to evaluate the discourse structure of non-native spontaneous speech based on these annotations; the highest performing feature showed a correlation of 0.612 with scores of discourse coherence provided by expert human raters.
Cite
CITATION STYLE
Wang, X., Bruno, J. V., Molloy, H. R., Evanini, K., & Zechner, K. (2017). Discourse annotation of non-native spontaneous spoken responses using the rhetorical structure theory framework. In ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 2, pp. 263–268). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/P17-2041
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.