🤖 AI Summary
Understanding the organizational principles and dynamic properties underlying large language models’ (LLMs) internal cognitive mechanisms remains a fundamental challenge. Method: We propose a modular community-detection–based cognitive network analysis paradigm, integrating cognitive science theory with plasticity-aware simulation to establish a cross-layer framework linking LLM parameters, skills, and architectural topology. Contribution/Results: We discover that LLM cognitive modules exhibit a distributed, brain-like organization—reminiscent of avian and small-mammalian neocortical architectures—where skill acquisition relies critically on dynamic, inter-regional coordination rather than static functional segregation. Empirical evaluation demonstrates that dynamic collaborative learning substantially outperforms rigid module-level interventions, providing both interpretable theoretical foundations and practical guidance for efficient fine-tuning. This work constitutes the first systematic investigation revealing biologically inspired organizational principles and emergent mechanisms governing LLM cognition.
📝 Abstract
Large Language Models (LLMs) have reshaped our world with significant advancements in science, engineering, and society through applications ranging from scientific discoveries and medical diagnostics to Chatbots. Despite their ubiquity and utility, the underlying mechanisms of LLM remain concealed within billions of parameters and complex structures, making their inner architecture and cognitive processes challenging to comprehend. We address this gap by adopting approaches to understanding emerging cognition in biology and developing a network-based framework that links cognitive skills, LLM architectures, and datasets, ushering in a paradigm shift in foundation model analysis. The skill distribution in the module communities demonstrates that while LLMs do not strictly parallel the focalized specialization observed in specific biological systems, they exhibit unique communities of modules whose emergent skill patterns partially mirror the distributed yet interconnected cognitive organization seen in avian and small mammalian brains. Our numerical results highlight a key divergence from biological systems to LLMs, where skill acquisition benefits substantially from dynamic, cross-regional interactions and neural plasticity. By integrating cognitive science principles with machine learning, our framework provides new insights into LLM interpretability and suggests that effective fine-tuning strategies should leverage distributed learning dynamics rather than rigid modular interventions.