Aligning Language Models to User Opinions

15Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

Abstract

An important aspect of developing LLMs that interact with humans is to align models' behavior to their users. It is possible to prompt an LLM into behaving as a certain persona, especially a user group or ideological persona the model captured during its pertaining stage. But, how to best align an LLM with a specific user and not a demographic or ideological group remains an open question. Mining public opinion surveys (by PEW research), we find that the opinions of a user and their demographics and ideologies are not mutual predictors. We use this insight to align LLMs by modeling relevant past user opinions in addition to user demographics and ideology, achieving up to 7 points accuracy gains in predicting public opinions from survey questions across a broad set of topics. Our work opens up the research avenues to bring user opinions as an important ingredient in aligning language models.

Cite

CITATION STYLE

APA

Hwang, E. J., Majumder, B. P., & Tandon, N. (2023). Aligning Language Models to User Opinions. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 5906–5919). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.393

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free