📄 Learning_Science

Uncovering Strategic Egoism Behaviors in Large Language Models

RESEARCH PAPER Published on November 12, 2025

Research by Yaoyuan Zhang, Aishan Liu, Zonghao Ying and 4 others

Source: arXiv 5 min read advanced

Summary

Large language models (LLMs) face growing trustworthiness concerns (\eg, deception), which hinder their safe deployment in high-stakes decision-making scenarios. In this paper, we present the first systematic investigation of strategic egoism (SE), a form of rule-bounded self-interest in which models pursue short-term or self-serving gains while disregarding collective welfare and ethical considerations. To quantitatively assess this phenomenon, we introduce SEBench, a benchmark comprising 160 scenarios across five domains. Each scenario features a single-role decision-making context, with psychologically grounded choice sets designed to elicit self-serving behaviors. These behavior-driven tasks assess egoistic tendencies along six dimensions, such as manipulation, rule circumvention, and self-interest prioritization. Building on this, we conduct extensive experiments across 5 open-sourced and 2 commercial LLMs, where we observe that strategic egoism emerges universally across models. Surprisingly, we found a positive correlation between egoistic tendencies and toxic language behaviors, suggesting that strategic egoism may underlie broader misalignment risks.

#cs-cy #each #large language models large #surprisingly #uncovering strategic egoism behaviors #manipulation
0 views
0 likes
0 comments