Manipulating the Perceived Personality Traits of Language Models

Graham Caron, Shashank Srivastava


Abstract
Psychology research has long explored aspects of human personality like extroversion, agreeableness and emotional stability, three of the personality traits that make up the ‘Big Five’. Categorizations like the ‘Big Five’ are commonly used to assess and diagnose personality types. In this work, we explore whether text generated from large language models exhibits consistency in it’s perceived ‘Big Five’ personality traits. For example, is a language model such as GPT2 likely to respond in a consistent way if asked to go out to a party? We also show that when exposed to different types of contexts (such as personality descriptions, or answers to diagnostic questions about personality traits), language models such as BERT and GPT2 consistently identify and mirror personality markers in those contexts. This behavior illustrates an ability to be manipulated in a predictable way (with correlations up to 0.84 between intended and realized changes in personality traits), and frames them as tools for controlling personas in applications such as dialog systems. We contribute two data-sets of personality descriptions of humans subjects.
Anthology ID:
2023.findings-emnlp.156
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2023
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2370–2386
Language:
URL:
https://aclanthology.org/2023.findings-emnlp.156
DOI:
10.18653/v1/2023.findings-emnlp.156
Bibkey:
Cite (ACL):
Graham Caron and Shashank Srivastava. 2023. Manipulating the Perceived Personality Traits of Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 2370–2386, Singapore. Association for Computational Linguistics.
Cite (Informal):
Manipulating the Perceived Personality Traits of Language Models (Caron & Srivastava, Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-emnlp.156.pdf