🤖 AI Summary
This study systematically quantifies gender bias against political figures on Reddit, analyzing 10 million politically themed comments. It proposes and empirically validates a five-dimensional bias framework—coverage, intersectionality, reference, sentiment, and lexical choice—to multidimensionally deconstruct explicit discrimination, benevolent sexism, and nonverbal cues in large-scale social media discourse for the first time. Methodologically, it integrates computational linguistic analysis, a custom bias classification schema, and fine-tuned models for sentiment and referential expression identification, complemented by cross-ideological subreddit comparisons. Results indicate that while female politicians receive comparable visibility to their male counterparts, their professional competence and authority are consistently linguistically attenuated—evidenced by higher rates of first-name usage, appearance- or family-related associations, and lower rates of title- or policy-based references. Bias is most pronounced in far-right communities but pervasive across both left- and right-wing subreddits. The study also releases a high-quality annotated dataset, establishing a new benchmark for research in political communication and algorithmic fairness.
📝 Abstract
Despite attempts to increase gender parity in politics, global efforts have struggled to ensure equal female representation. This is likely tied to implicit gender biases against women in authority. In this work, we present a comprehensive study of gender biases that appear in online political discussion. To this end, we collect 10 million comments on Reddit in conversations about male and female politicians, which enables an exhaustive study of automatic gender bias detection. We address not only misogynistic language, but also other manifestations of bias, like benevolent sexism in the form of seemingly positive sentiment and dominance attributed to female politicians, or differences in descriptor attribution. Finally, we conduct a multi-faceted study of gender bias towards politicians investigating both linguistic and extra-linguistic cues. We assess 5 different types of gender bias, evaluating coverage, combinatorial, nominal, sentimental and lexical biases extant in social media language and discourse. Overall, we find that, contrary to previous research, coverage and sentiment biases suggest equal public interest in female politicians. Rather than overt hostile or benevolent sexism, the results of the nominal and lexical analyses suggest this interest is not as professional or respectful as that expressed about male politicians. Female politicians are often named by their first names and are described in relation to their body, clothing, or family; this is a treatment that is not similarly extended to men. On the now banned far-right subreddits, this disparity is greatest, though differences in gender biases still appear in the right and left-leaning subreddits. We release the curated dataset to the public for future studies.