π€ AI Summary
This study investigates the capability and controllability of large language models (LLMs) in simulating emotions and personality traits during social interaction. To this end, we introduce PsySETβthe first comprehensive benchmark for systematically evaluating diverse LLMs and steering strategies (prompt engineering, fine-tuning, and representation-space vector injection) under emotion- and personality-conditioned guidance. Our key findings reveal nonlinear side effects of affective steering: e.g., induced joy degrades factual robustness, while anger enhances privacy resistance. We propose the first multidimensional credibility metric spanning safety, authenticity, fairness, and ethics, uncovering heterogeneous impacts of psychological states on safety risks and bias propagation. Empirical results show that prompt-based steering offers usability but lacks granularity, whereas vector injection enables fine-grained control at a modest cost to output quality.
π Abstract
The ability to control LLMs' emulated emotional states and personality traits is essential for enabling rich, human-centered interactions in socially interactive settings. We introduce PsySET, a Psychologically-informed benchmark to evaluate LLM Steering Effectiveness and Trustworthiness across the emotion and personality domains. Our study spans four models from different LLM families paired with various steering strategies, including prompting, fine-tuning, and representation engineering. Our results indicate that prompting is consistently effective but limited in intensity control, whereas vector injections achieve finer controllability while slightly reducing output quality. Moreover, we explore the trustworthiness of steered LLMs by assessing safety, truthfulness, fairness, and ethics, highlighting potential side effects and behavioral shifts. Notably, we observe idiosyncratic effects; for instance, even a positive emotion like joy can degrade robustness to adversarial factuality, lower privacy awareness, and increase preferential bias. Meanwhile, anger predictably elevates toxicity yet strengthens leakage resistance. Our framework establishes the first holistic evaluation of emotion and personality steering, offering insights into its interpretability and reliability for socially interactive applications.