🤖 AI Summary
This work addresses the structural limitations of prevailing alignment approaches—such as value flattening, inadequate normative representation, and epistemic blind spots—when confronted with pluralistic values, multiple stakeholders, and irreducible uncertainties. It proposes a novel “marginal alignment” paradigm that transcends the conventional scalar reward framework by integrating multidimensional value modeling, democratic representational mechanisms, and cognitively interactive design, thereby reconceptualizing alignment as a dynamic, full-cycle process of normative governance. The study develops a comprehensive three-phase, seven-pillar framework spanning data collection, training, and evaluation to systematically expose the bottlenecks of generic alignment methods and offers an actionable pathway that harmonizes technical implementation with institutional coordination, opening new directions for AI alignment in complex sociotechnical systems.
📝 Abstract
Large language models are being deployed in complex socio-technical systems, which exposes limits in current alignment practice. We take the position that the dominant paradigm of General Alignment, which compresses diverse human values into a single scalar reward, reaches a structural ceiling in settings with conflicting values, plural stakeholders, and irreducible uncertainty. These failures follow from the mathematics and incentives of scalarization and lead to \textbf{structural} value flattening, \textbf{normative} representation loss, and \textbf{cognitive} uncertainty blindness. We introduce Edge Alignment as a distinct approach in which systems preserve multi dimensional value structure, support plural and democratic representation, and incorporate epistemic mechanisms for interaction and clarification. To make this approach practical, we propose seven interdependent pillars organized into three phases. We identify key challenges in data collection, training objectives, and evaluation, outlining complementary technical and governance directions. Taken together, these measures reframe alignment as a lifecycle problem of dynamic normative governance rather than as a single instance optimization task.