đ€ AI Summary
This work investigates how large language models (LLMs) model annotator personas across the data perspectivism spectrumâfrom strong personalization to weak aggregationâto detect hate speech and offensive content. Method: We introduce demographically grounded, predefined annotator personas and integrate them with LLM-generated annotations and established perspective modeling techniques to systematically evaluate annotation consistency and human alignment under varying perspective strengths. Results: LLMs selectively leverage demographic cues in personas to produce annotations biased toward group consensus; they significantly outperform both conventional methods and human annotators under weak-perspective settings, while approachingâthough not surpassingâhuman performance under strong-perspective conditions, revealing a âdepersonalized aggregationâ tendency. This study is the first to systematically incorporate data perspectivism into LLM-based content annotation, establishing a novel paradigm for interpretable and controllable modeling of subjectivity.
đ Abstract
In this work, we explore the capability of Large Language Models (LLMs) to annotate hate speech and abusiveness while considering predefined annotator personas within the strong-to-weak data perspectivism spectra. We evaluated LLM-generated annotations against existing annotator modeling techniques for perspective modeling. Our findings show that LLMs selectively use demographic attributes from the personas. We identified prototypical annotators, with persona features that show varying degrees of alignment with the original human annotators. Within the data perspectivism paradigm, annotator modeling techniques that do not explicitly rely on annotator information performed better under weak data perspectivism compared to both strong data perspectivism and human annotations, suggesting LLM-generated views tend towards aggregation despite subjective prompting. However, for more personalized datasets tailored to strong perspectivism, the performance of LLM annotator modeling approached, but did not exceed, human annotators.