🤖 AI Summary
In health-related online discussions, user toxicity frequently triggers social conflict and promotes pseudoscientific behavior; conventional post-hoc detection-and-removal strategies often backfire. This paper introduces a “predictive intervention” paradigm—first applying collaborative filtering to model cross-user–subcommunity toxic interaction likelihood, enabling pre-participation prediction of whether a given user will post toxic content in a specific health subcommunity (e.g., COVID-19–related Reddit forums), without relying on reactive moderation. Our method jointly encodes user behavioral representations and subcommunity-level semantic features to enable fine-grained modeling of toxicity compatibility. Evaluated on real-world COVID-19 Reddit data, the model achieves AUC and F1 scores exceeding 80%, substantially enhancing decision support for conflict avoidance. Core contributions include: (1) establishing the first predictive framework for toxicity interactions, and (2) shifting toxicity governance from reactive mitigation to proactive prevention.
📝 Abstract
In health-related topics, user toxicity in online discussions frequently becomes a source of social conflict or promotion of dangerous, unscientific behaviour; common approaches for battling it include different forms of detection, flagging and/or removal of existing toxic comments, which is often counterproductive for platforms and users alike. In this work, we propose the alternative of combatting user toxicity predictively, anticipating where a user could interact toxically in health-related online discussions. Applying a Collaborative Filtering-based Machine Learning methodology, we predict the toxicity in COVID-related conversations between any user and subcommunity of Reddit, surpassing 80% predictive performance in relevant metrics, and allowing us to prevent the pairing of conflicting users and subcommunities.