๐ค AI Summary
This study addresses the widespread lack of personalized adaptation in hate speech moderation mechanisms across different interface contexts on social media platforms. Through a three-round Delphi study involving 40 activists who have experienced online hate speech, the research systematically investigates user preferences for customizable moderation tools across distinct Instagram interfacesโsuch as comments, Reels, and the main feed. Integrating quantitative ratings, rankings, and qualitative feedback, the analysis focuses on dimensions including input modality, content type, and degree of automation. The findings reveal, for the first time, divergent user preferences between conversational and algorithmically recommended interface contexts. Emphasizing the importance of cross-screen reversibility and user oversight, the study concludes with user-centered design recommendations for personalized content moderation systems.
๐ Abstract
Hate speech remains a pressing challenge on social media, where platform moderation often fails to protect targeted users. Personal moderation tools that let users decide how content is filtered can address some of these shortcomings. However, it remains an open question on which screens (e.g., the comments, the reels tab, or the home feed) users want personal moderation and which features they value most. To address these gaps, we conducted a three-wave Delphi study with 40 activists who experienced hate speech. We combined quantitative ratings and rankings with open questions about required features. Participants prioritized personal moderation for conversational and algorithmically curated screens. They valued features allowing for reversibility and oversight across screens, while input-based, content-type specific, and highly automated features are more screen specific. We discuss the importance of personal moderation and offer user-centered design recommendations for personal moderation on Instagram.