๐ค AI Summary
Existing LLM-based social simulations fail to capture the feedback loop between individual decision-making and collective dynamics, leading to significant deviations from real-world behavioral data. To address this, we propose Mean-Field LLMโa novel framework that pioneers the integration of mean-field theory into LLM-driven social simulation. It explicitly models bidirectional microโmacro feedback via alternating rollouts between a strategic agent model and a mean-field population model. We further introduce IB-Tune, an information-bottleneck-driven fine-tuning method, to enhance predictive accuracy of population-level distributions while promoting representation parsimony. The framework adopts a domain-agnostic architecture for cross-domain generalization. Evaluated on real-world social datasets, it reduces KL divergence by 47% over baselines, substantially improving trend forecasting and intervention planning capabilities. It generalizes robustly across seven distinct domains and four major LLM backbones.
๐ Abstract
Simulating collective decision-making involves more than aggregating individual behaviors; it arises from dynamic interactions among individuals. While large language models (LLMs) show promise for social simulation, existing approaches often exhibit deviations from real-world data. To address this gap, we propose the Mean-Field LLM (MF-LLM) framework, which explicitly models the feedback loop between micro-level decisions and macro-level population. MF-LLM alternates between two models: a policy model that generates individual actions based on personal states and group-level information, and a mean field model that updates the population distribution from the latest individual decisions. Together, they produce rollouts that simulate the evolving trajectories of collective decision-making. To better match real-world data, we introduce IB-Tune, a fine-tuning method for LLMs grounded in the information bottleneck principle, which maximizes the relevance of population distributions to future actions while minimizing redundancy with historical data. We evaluate MF-LLM on a real-world social dataset, where it reduces KL divergence to human population distributions by 47 percent over non-mean-field baselines, and enables accurate trend forecasting and intervention planning. It generalizes across seven domains and four LLM backbones, providing a scalable foundation for high-fidelity social simulation.