Moral Susceptibility and Robustness under Persona Role-Play in Large Language Models
NeutralArtificial Intelligence
The study on moral responses of large language models (LLMs) under persona role-play sheds light on how these models navigate moral judgments in social contexts. By employing the Moral Foundations Questionnaire (MFQ), researchers established a benchmark to quantify moral susceptibility and robustness. Results indicated that the Claude model is the most robust, significantly outperforming Gemini and GPT-4, while larger variants of models tend to be more susceptible to moral shifts. This suggests that the model family plays a crucial role in determining moral robustness, as it accounts for most of the variance observed. Furthermore, the positive correlation between robustness and susceptibility at the family level emphasizes the intricate dynamics of LLMs, which are increasingly relevant as these technologies integrate into societal frameworks.
— via World Pulse Now AI Editorial System

