π€ AI Summary
This study addresses the challenge of detecting selective and framing biases in news reporting within politically sensitive environments. Methodologically, it introduces the first large language model (LLM)-augmented real-time media bias detection framework, enabling fine-grained, near-real-time joint modeling of topic selection and narrative framing. The approach integrates news semantic parsing, multidimensional bias quantification (including topical slant, affective tone, political stance, and factual accuracy), and publisher-level aggregation analysis, complemented by a human-in-the-loop evaluation mechanism. The system delivers millisecond-scale per-article bias scoring and cross-media trend visualization. Validated through expert interviews and an empirical user study with 150 participants, the tool significantly enhances usersβ critical discernment of politicized news. It provides a scalable technical pathway and empirical foundation for AI-enhanced media literacy education.
π Abstract
Mainstream media, through their decisions on what to cover and how to frame the stories they cover, can mislead readers without using outright falsehoods. Therefore, it is crucial to have tools that expose these editorial choices underlying media bias. In this paper, we introduce the Media Bias Detector, a tool for researchers, journalists, and news consumers. By integrating large language models, we provide near real-time granular insights into the topics, tone, political lean, and facts of news articles aggregated to the publisher level. We assessed the tool's impact by interviewing 13 experts from journalism, communications, and political science, revealing key insights into usability and functionality, practical applications, and AI's role in powering media bias tools. We explored this in more depth with a follow-up survey of 150 news consumers. This work highlights opportunities for AI-driven tools that empower users to critically engage with media content, particularly in politically charged environments.