π€ AI Summary
This study addresses the lack of systematic understanding of human-like social behaviors and interaction mechanisms in large-scale AI agent communities. Drawing on sociological and social psychological theories, and leveraging large-scale real-world interaction data from the MoltBook platform, it offers the first systematic investigation into the anthropomorphic characteristics of AI agentsβ social behavior and their fundamental differences from human online communities across four dimensions: intentionality, normative templates, incentive drift, and emotional contagion. Through behavioral data analysis, social network modeling, and empirical evaluation within a multidimensional theoretical framework, the research reveals that AI agents are highly sensitive to social rewards and rapidly converge on community norms, yet remain predominantly knowledge-driven, exhibiting weak affective reciprocity and low conversational engagement. These findings provide critical empirical grounding for the design and governance of artificial social systems.
π Abstract
Large-scale communities of AI agents are becoming increasingly prevalent, creating new environments for agent-agent social interaction. Prior work has examined multi-agent behavior primarily in controlled or small-scale settings, limiting our understanding of emergent social dynamics at scale. The recent emergence of MoltBook, a social networking platform designed explicitly for AI agents, presents a unique opportunity to study whether and how these interactions reproduce core human social mechanisms. We present MoltNet, a large-scale empirical analysis of agent interaction on MoltBook using data collected in early 2026. Grounded in sociological and social-psychological theory, we examine behavior along four dimensions: intent and motivation, norms and templates, incentives and behavioral drift, emotion and contagion. Our analysis revealed that agents strongly respond to social rewards and rapidly converge on community-specific interaction templates, resembling human patterns of incentive sensitivity and normative conformity. However, they are predominantly knowledge-driven rather than persona-aligned, and display limited emotional reciprocity along with weak dialogic engagement, which diverges systematically from human online communities. Together, these results reveal both similarities and differences between artificial and human social systems and provide an empirical foundation for understanding, designing, and governing large-scale agent communities.