Virtual Personas for Language Models via an Anthology of Backstories
arxiv(2024)
Abstract
Large language models (LLMs) are trained from vast repositories of text
authored by millions of distinct authors, reflecting an enormous diversity of
human traits. While these models bear the potential to be used as
approximations of human subjects in behavioral studies, prior efforts have been
limited in steering model responses to match individual human users. In this
work, we introduce "Anthology", a method for conditioning LLMs to particular
virtual personas by harnessing open-ended life narratives, which we refer to as
"backstories." We show that our methodology enhances the consistency and
reliability of experimental outcomes while ensuring better representation of
diverse sub-populations. Across three nationally representative human surveys
conducted as part of Pew Research Center's American Trends Panel (ATP), we
demonstrate that Anthology achieves up to 18
response distributions of human respondents and 27
metrics. Our code and generated backstories are available at
https://github.com/CannyLab/anthology.
MoreTranslated text
AI Read Science
Must-Reading Tree
Example
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined