🤖 AI Summary
Current large language models (LLMs) lack ethical consistency across contexts and temporal scales, while mainstream alignment methods rely on static datasets and post-hoc evaluation, impeding dynamic monitoring. To address this, we propose the Moral Consistency Pipeline (MoCoP)—the first data-free, model-agnostic, closed-loop ethical assessment framework. MoCoP integrates lexical completeness analysis, semantic risk estimation, and reasoning-driven judgment modeling to recast ethical evaluation as an autonomous introspective process. Its self-generating–self-evaluating–self-optimizing architecture enables reproducible, scalable, continuous auditing. Experiments on GPT-4-Turbo and DeepSeek demonstrate a strong negative correlation between ethical stability and toxicity (r = −0.81, p < 0.001), with near-zero correlation to response latency (r ≈ 0), confirming its potential as an intrinsic model stability property. This advances computational ethics toward dynamic, introspective paradigms.
📝 Abstract
The rapid advancement and adaptability of Large Language Models (LLMs) highlight the need for moral consistency, the capacity to maintain ethically coherent reasoning across varied contexts. Existing alignment frameworks, structured approaches designed to align model behavior with human ethical and social norms, often rely on static datasets and post-hoc evaluations, offering limited insight into how ethical reasoning may evolve across different contexts or temporal scales. This study presents the Moral Consistency Pipeline (MoCoP), a dataset-free, closed-loop framework for continuously evaluating and interpreting the moral stability of LLMs. MoCoP combines three supporting layers: (i) lexical integrity analysis, (ii) semantic risk estimation, and (iii) reasoning-based judgment modeling within a self-sustaining architecture that autonomously generates, evaluates, and refines ethical scenarios without external supervision. Our empirical results on GPT-4-Turbo and DeepSeek suggest that MoCoP effectively captures longitudinal ethical behavior, revealing a strong inverse relationship between ethical and toxicity dimensions (correlation rET = -0.81, p value less than 0.001) and a near-zero association with response latency (correlation rEL approximately equal to 0). These findings demonstrate that moral coherence and linguistic safety tend to emerge as stable and interpretable characteristics of model behavior rather than short-term fluctuations. Furthermore, by reframing ethical evaluation as a dynamic, model-agnostic form of moral introspection, MoCoP offers a reproducible foundation for scalable, continuous auditing and advances the study of computational morality in autonomous AI systems.