Explainable Ethical Assessment on Human Behaviors by Generating Conflicting Social Norms

📅 2025-12-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current AI systems exhibit low interpretability and credibility in ethical evaluation of human behavior due to the absence of explicit modeling of social norms. This paper addresses the moral trade-off challenge arising from coexisting conflicting social norms (e.g., courage vs. self-preservation) by proposing ClarityEthic—the first framework to formalize normative conflict as a core mechanism in ethical reasoning. Methodologically, it integrates large language model–driven norm generation, social norm embedding, contrastive learning–enhanced moral attribution, and multi-dimensional value orientation prediction. ClarityEthic achieves significant improvements over strong baselines across multiple ethical evaluation benchmarks. Human evaluations confirm that the generated norms are highly plausible and explanatory, substantially enhancing decision transparency and model trustworthiness.

Technology Category

Application Category

📝 Abstract
Human behaviors are often guided or constrained by social norms, which are defined as shared, commonsense rules. For example, underlying an action `` extit{report a witnessed crime}" are social norms that inform our conduct, such as `` extit{It is expected to be brave to report crimes}''. Current AI systems that assess valence (i.e., support or oppose) of human actions by leveraging large-scale data training not grounded on explicit norms may be difficult to explain, and thus untrustworthy. Emulating human assessors by considering social norms can help AI models better understand and predict valence. While multiple norms come into play, conflicting norms can create tension and directly influence human behavior. For example, when deciding whether to `` extit{report a witnessed crime}'', one may balance extit{bravery} against extit{self-protection}. In this paper, we introduce extit{ClarityEthic}, a novel ethical assessment approach, to enhance valence prediction and explanation by generating conflicting social norms behind human actions, which strengthens the moral reasoning capabilities of language models by using a contrastive learning strategy. Extensive experiments demonstrate that our method outperforms strong baseline approaches, and human evaluations confirm that the generated social norms provide plausible explanations for the assessment of human behaviors.
Problem

Research questions and friction points this paper is trying to address.

Enhance AI ethical assessment with explainable conflicting social norms
Improve valence prediction of human behaviors using moral reasoning
Generate plausible explanations for ethical evaluations via contrastive learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Generates conflicting social norms for ethical assessment
Uses contrastive learning to strengthen moral reasoning
Enhances valence prediction with explainable norm-based explanations
🔎 Similar Papers
No similar papers found.
Yuxi Sun
Yuxi Sun
Shenzhen University
Remote Sensing Image RetrievalVisual Grounding
W
Wei Gao
School of Computing and Information Systems, Singapore Management University
Hongzhan Lin
Hongzhan Lin
Hong Kong Baptist University
Natural Language ProcessingMultimodal ReasoningSocial Computing
J
Jing Ma
Department of Computer Science, Hong Kong Baptist University
W
Wenxuan Zhang
Information Systems Technology and Design, Singapore University of Technology and Design