Discrepancy Detection at the Data Level: Toward Consistent Multilingual Question Answering

📅 2025-10-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Multilingual question-answering (QA) systems face dual challenges of objective factual consistency and subjective cultural adaptability—particularly critical in sensitive domains such as maternal and child health. To address this, we propose MIND, the first framework to integrate user-coordinated verification into multilingual factual consistency detection, explicitly distinguishing factual errors from culturally appropriate variations. Our method combines multilingual NLP techniques, context-aware answer comparison, culture-sensitive annotation, and bilingual human validation. Leveraging this pipeline, we construct the first publicly available bilingual dataset annotated for fact–culture inconsistency. Experiments demonstrate that MIND effectively identifies cross-lingual divergent answers across multiple domains, exhibits strong cross-domain generalization, and provides a scalable technical pathway—along with benchmark resources—for developing culturally aware, factually reliable multilingual QA systems.

Technology Category

Application Category

📝 Abstract
Multilingual question answering (QA) systems must ensure factual consistency across languages, especially for objective queries such as What is jaundice?, while also accounting for cultural variation in subjective responses. We propose MIND, a user-in-the-loop fact-checking pipeline to detect factual and cultural discrepancies in multilingual QA knowledge bases. MIND highlights divergent answers to culturally sensitive questions (e.g., Who assists in childbirth?) that vary by region and context. We evaluate MIND on a bilingual QA system in the maternal and infant health domain and release a dataset of bilingual questions annotated for factual and cultural inconsistencies. We further test MIND on datasets from other domains to assess generalization. In all cases, MIND reliably identifies inconsistencies, supporting the development of more culturally aware and factually consistent QA systems.
Problem

Research questions and friction points this paper is trying to address.

Detecting factual inconsistencies in multilingual QA systems
Identifying cultural variations in subjective question responses
Ensuring cross-lingual consistency for objective factual queries
Innovation

Methods, ideas, or system contributions that make the work stand out.

User-in-the-loop fact-checking pipeline for discrepancies
Detects factual and cultural inconsistencies in multilingual QA
Evaluated generalization across domains for reliable performance
🔎 Similar Papers
No similar papers found.
L
Lorena Calvo-Bartolomé
Universidad Carlos III de Madrid, Spain
V
Valérie Aldana
University of Maryland, College Park, USA
K
Karla Cantarero
University of Maryland, College Park, USA
A
Alonso Madroñal de Mesa
Universidad Carlos III de Madrid, Spain
Jerónimo Arenas-García
Jerónimo Arenas-García
Universidad Carlos III de Madrid, Spain
Jordan Boyd-Graber
Jordan Boyd-Graber
Full Professor, University of Maryland
Machine LearningComputational LinguisticsNatural Language ProcessingTopic ModelsQuestion Answering