đ¤ AI Summary
This study investigates whether proprietary large language models (LLMs) implicitly sanitize sensitive contentâeven without explicit instructions or fine-tuningârevealing potential intrinsic value alignment.
Method: Using GPT-4o-mini as a testbed, we conduct zero-shot sensitivity classification and empirical paraphrasing experiments to systematically quantify the degree to which the model attenuates derogatory and taboo language during rewriting.
Contribution/Results: We provide the first empirical evidence that LLMs exhibit inherent content moderation tendencies despite lacking domain-specific sensitive-content training: rewritten outputs show statistically significant reductions in sensitivity scores and marked decreases in derogatory/taboo lexical usage. Moreover, the modelâs zero-shot sensitivity classification accuracy surpasses that of conventional baseline models. These findings uncover latent value-alignment mechanisms in black-box LLMs, offering novel insights into implicit safety boundaries and the embeddedness of normative values in generative AI systems.
đ Abstract
Proprietary Large Language Models (LLMs) have shown tendencies toward politeness, formality, and implicit content moderation. While previous research has primarily focused on explicitly training models to moderate and detoxify sensitive content, there has been limited exploration of whether LLMs implicitly sanitize language without explicit instructions. This study empirically analyzes the implicit moderation behavior of GPT-4o-mini when paraphrasing sensitive content and evaluates the extent of sensitivity shifts. Our experiments indicate that GPT-4o-mini systematically moderates content toward less sensitive classes, with substantial reductions in derogatory and taboo language. Also, we evaluate the zero-shot capabilities of LLMs in classifying sentence sensitivity, comparing their performances against traditional methods.