๐ค AI Summary
This work addresses the lack of a unified theoretical foundation for belief representation in large language models (LLMs), which leads to inconsistency and opacity in their internal belief states. We propose the first formalized standard framework for belief representation, defining three core properties: semantic consistency, dynamic updatability, and causal traceability. Methodologically, we integrate doxastic logic (DoX), neuro-symbolic interfaces, inter-layer attention attribution, and counterfactual belief editing to enable verifiable modeling and controllable intervention of LLMsโ latent belief states. Evaluated on BELIEF-BENCH, our approach improves belief consistency accuracy by 32.7% and supports fine-grained belief injection and withdrawal. The framework establishes the first auditable, cross-task generalizable cognitive infrastructure for trustworthy AI, effectively breaking the โbelief black boxโ limitation of LLMs.