How Implicit Bias Accumulates and Propagates in LLM Long-term Memory

📅 2026-02-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the risk that large language models may implicitly accumulate and propagate biases across domains through long-term memory mechanisms, raising significant fairness concerns. To systematically investigate this issue, the authors introduce the Decision-based Implicit Bias (DIB) benchmark and a long-horizon interactive simulation framework to evaluate bias dynamics in mainstream models and memory architectures. Their analysis reveals, for the first time, that implicit bias exhibits non-stationary growth within long-term memory and can transfer across domains. To mitigate this, they propose Dynamic Memory Tagging (DMT), a mechanism that enforces fairness constraints during the memory writing phase. Experimental results demonstrate that DMT effectively suppresses bias accumulation and blocks cross-domain propagation, achieving more persistent and robust debiasing performance compared to static prompting strategies.

Technology Category

Application Category

📝 Abstract
Long-term memory mechanisms enable Large Language Models (LLMs) to maintain continuity and personalization across extended interaction lifecycles, but they also introduce new and underexplored risks related to fairness. In this work, we study how implicit bias, defined as subtle statistical prejudice, accumulates and propagates within LLMs equipped with long-term memory. To support systematic analysis, we introduce the Decision-based Implicit Bias (DIB) Benchmark, a large-scale dataset comprising 3,776 decision-making scenarios across nine social domains, designed to quantify implicit bias in long-term decision processes. Using a realistic long-horizon simulation framework, we evaluate six state-of-the-art LLMs integrated with three representative memory architectures on DIB and demonstrate that LLMs'implicit bias does not remain static but intensifies over time and propagates across unrelated domains. We further analyze mitigation strategies and show that a static system-level prompting baseline provides limited and short-lived debiasing effects. To address this limitation, we propose Dynamic Memory Tagging (DMT), an agentic intervention that enforces fairness constraints at memory write time. Extensive experimental results show that DMT substantially reduces bias accumulation and effectively curtails cross-domain bias propagation.
Problem

Research questions and friction points this paper is trying to address.

implicit bias
long-term memory
fairness
bias propagation
large language models
Innovation

Methods, ideas, or system contributions that make the work stand out.

implicit bias
long-term memory
Dynamic Memory Tagging
fairness
LLM
🔎 Similar Papers
No similar papers found.
Y
Yiming Ma
Chongqing Research Institute of Harbin Institute of Technology, Chongqing, China
Lixu Wang
Lixu Wang
Northwestern University
Machine LearningData Privacy
L
Lionel Z. Wang
Department of Management and Marketing, The Hong Kong Polytechnic University, Hong Kong
H
Hongkun Yang
Haide College, Ocean University of China, Qingdao, China
H
Haoming Sun
School of Computer Science, The University of Sheffield, Sheffield, UK
Xin Xu
Xin Xu
The Hong Kong Polytechnic University
Digital TransformationHuman-AI InteractionBusiness Analytics
J
Jiaqi Wu
Department of Automation, Tsinghua University, Beijing, China
B
Bin Chen
Chongqing Research Institute of Harbin Institute of Technology, Chongqing, China
Wei Dong
Wei Dong
PHD candidate, School of Computer Science and Engineering, Northwestern Polytechnical University,
Deep Learning