Reading StackOverflow Encourages Cheating: Adding Question Text Improves Extractive Code Generation

4Citations
Citations of this article
63Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Answering a programming question using only its title is difficult as salient contextual information is omitted. Based on this observation, we present a corpus of over 40, 000 StackOverflow question texts to be used in conjunction with their corresponding intents from the CoNaLa dataset (Yin et al., 2018). Using both the intent and question body, we use BART to establish a baseline BLEU score of 34.35 for this new task. We find further improvements of 2.8% by combining the mined CoNaLa data with the labeled data to achieve a 35.32 BLEU score. We evaluate prior state-of-the-art CoNaLa models with this additional data and find that our proposed method of using the body and mined data beats the BLEU score of the prior state-of-the-art by 71.96%. Finally, we perform ablations to demonstrate that BART is an unsupervised multimodal learner and examine its extractive behavior.

Cite

CITATION STYLE

APA

Orlanski, G., & Gittens, A. (2021). Reading StackOverflow Encourages Cheating: Adding Question Text Improves Extractive Code Generation. In NLP4Prog 2021 - 1st Workshop on Natural Language Processing for Programming, Proceedings of the Workshop (pp. 65–76). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.nlp4prog-1.8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free