Large Language Models are Biased Because They Are Large Language Models

📅 2024-06-19
🏛️ arXiv.org
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
This paper investigates the structural origins of harmful biases in large language models (LLMs), arguing that such biases are not artifacts of data noise or training errors, but rather inherent deficiencies arising from the coupling of scale expansion, the statistical nature of language modeling, and the data-driven paradigm. Method: Through rigorous theoretical analysis, derivation from first principles of language modeling, and critical reflection on mechanisms of socio-semantic embedding, the authors systematically demonstrate—novelly—that bias is fundamentally irreducible: it cannot be fully eliminated via post-hoc mitigation, dataset curation, or alignment fine-tuning. Contribution/Results: The core contribution is the original thesis that “bias is an architectural problem,” necessitating a fundamental reexamination of LLM design assumptions. This reframes AI ethics discourse away from incremental technical fixes toward foundational paradigm shifts in model architecture and learning principles.

Technology Category

Application Category

📝 Abstract
This position paper's primary goal is to provoke thoughtful discussion about the relationship between bias and fundamental properties of large language models. I do this by seeking to convince the reader that harmful biases are an inevitable consequence arising from the design of any large language model as LLMs are currently formulated. To the extent that this is true, it suggests that the problem of harmful bias cannot be properly addressed without a serious reconsideration of AI driven by LLMs, going back to the foundational assumptions underlying their design.
Problem

Research questions and friction points this paper is trying to address.

Examines bias in large language models (LLMs).
Links harmful biases to LLM design principles.
Calls for rethinking AI foundations to address bias.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Examines bias in large language models
Links bias to LLM design fundamentals
Calls for rethinking AI foundational assumptions
🔎 Similar Papers
No similar papers found.