🤖 AI Summary
Existing hallucination evaluation datasets lack comprehensive coverage of fine-grained hallucinations—specifically at the entity, relation, and sentence levels—across multilingual settings. Method: This paper introduces HalluVerse25, the first fine-grained multilingual hallucination benchmark for English, Arabic, and Turkish. It proposes a novel paradigm combining LLM-controlled hallucination injection with collaborative annotation by trilingual domain experts, augmented by a cross-lingual consistency verification mechanism. Contribution/Results: HalluVerse25 enables the first-ever fine-grained, multilingual hallucination classification annotation, substantially improving sample authenticity and inter-annotator agreement. Evaluation on HalluVerse25 reveals that proprietary models consistently outperform open-source counterparts in hallucination detection, while all models exhibit the weakest performance on Arabic—highlighting a critical bottleneck in current multilingual hallucination identification capabilities.
📝 Abstract
Large Language Models (LLMs) are increasingly used in various contexts, yet remain prone to generating non-factual content, commonly referred to as"hallucinations". The literature categorizes hallucinations into several types, including entity-level, relation-level, and sentence-level hallucinations. However, existing hallucination datasets often fail to capture fine-grained hallucinations in multilingual settings. In this work, we introduce HalluVerse25, a multilingual LLM hallucination dataset that categorizes fine-grained hallucinations in English, Arabic, and Turkish. Our dataset construction pipeline uses an LLM to inject hallucinations into factual biographical sentences, followed by a rigorous human annotation process to ensure data quality. We evaluate several LLMs on HalluVerse25, providing valuable insights into how proprietary models perform in detecting LLM-generated hallucinations across different contexts.