On- And off-topic classification and semantic annotation of user-generated software requirements

19Citations
Citations of this article
95Readers
Mendeley users who have this article in their library.

Abstract

Users prefer natural language software requirements because of their usability and accessibility. When they describe their wishes for software development, they often provide off-topic information. We therefore present REaCT1, an automated approach for identifying and semantically annotating the on-topic parts of requirement descriptions. It is designed to support requirement engineers in the elicitation process on detecting and analyzing requirements in user-generated content. Since no lexical resources with domain-specific information about requirements are available, we created a corpus of requirements written in controlled language by instructed users and uncontrolled language by uninstructed users. We annotated these requirements regarding predicate-argument structures, conditions, priorities, motivations and semantic roles and used this information to train classifiers for information extraction purposes. REaCT achieves an accuracy of 92% for the on- and off-topic classification task and an F1measure of 72% for the semantic annotation.

Cite

CITATION STYLE

APA

Dollmann, M., & Geierhos, M. (2016). On- And off-topic classification and semantic annotation of user-generated software requirements. In EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1807–1816). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d16-1186

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free