Despite exciting progress in causal language models, the expressiveness of their representations is largely limited due to poor discrimination ability. To remedy this issue, we present CONTRACLM, a novel contrastive learning framework at both the token-level and the sequence-level. We assess CONTRACLM on a variety of downstream tasks. We show that CONTRACLM enhances the discrimination of representations and bridges the gap with encoder-only models, which makes causal language models better suited for tasks beyond language generation. Specifically, we attain 44% relative improvement on the Semantic Textual Similarity tasks and 34% on Code-to-Code Search tasks. Furthermore, by improving the expressiveness of representations, CONTRACLM also boosts the source code generation capability with 9% relative improvement on execution accuracy on the HumanEval benchmark..
CITATION STYLE
Jain, N., Zhang, D., Ahmad, W. U., Wang, Z., Nan, F., Li, X., … Xiang, B. (2023). CONTRACLM: Contrastive Learning For Causal Language Model. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 6436–6459). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.355
Mendeley helps you to discover research relevant for your work.