Phantom: Personality Has An Effect On Theory-of-mind Reasoning In Large Language Models · The Large Language Model Bible Contribute to LLM-Bible

Phantom: Personality Has An Effect On Theory-of-mind Reasoning In Large Language Models

Tan Fiona Anting, Yeo Gerard Christopher, Wu Fanyou, Xu Weijie, Jain Vinija, Chadha Aman, Jaidka Kokil, Liu Yang, Ng See-kiong. Arxiv 2024

[Paper]    
GPT Model Architecture Prompting

Recent advances in large language models (LLMs) demonstrate that their capabilities are comparable, or even superior, to humans in many tasks in natural language processing. Despite this progress, LLMs are still inadequate at social-cognitive reasoning, which humans are naturally good at. Drawing inspiration from psychological research on the links between certain personality traits and Theory-of-Mind (ToM) reasoning, and from prompt engineering research on the hyper-sensitivity of prompts in affecting LLMs capabilities, this study investigates how inducing personalities in LLMs using prompts affects their ToM reasoning capabilities. Our findings show that certain induced personalities can significantly affect the LLMs’ reasoning capabilities in three different ToM tasks. In particular, traits from the Dark Triad have a larger variable effect on LLMs like GPT-3.5, Llama 2, and Mistral across the different ToM tasks. We find that LLMs that exhibit a higher variance across personality prompts in ToM also tends to be more controllable in personality tests: personality traits in LLMs like GPT-3.5, Llama 2 and Mistral can be controllably adjusted through our personality prompts. In today’s landscape where role-play is a common strategy when using LLMs, our research highlights the need for caution, as models that adopt specific personas with personalities potentially also alter their reasoning abilities in an unexpected manner.

Similar Work