MultiHal: Multilingual Dataset for Knowledge-Graph Grounded Evaluation of LLM Hallucinations
Abstract
A multilingual, multihop benchmark using knowledge graphs for evaluating and mitigating hallucinations in large language models.
Large Language Models (LLMs) have inherent limitations of faithfulness and factuality, commonly referred to as hallucinations. Several benchmarks have been developed that provide a test bed for factuality evaluation within the context of English-centric datasets, while relying on supplementary informative context like web links or text passages but ignoring the available structured factual resources. To this end, Knowledge Graphs (KGs) have been identified as a useful aid for hallucination mitigation, as they provide a structured way to represent the facts about entities and their relations with minimal linguistic overhead. We bridge the lack of KG paths and multilinguality for factual language modeling within the existing hallucination evaluation benchmarks and propose a KG-based multilingual, multihop benchmark called MultiHal framed for generative text evaluation. As part of our data collection pipeline, we mined 140k KG-paths from open-domain KGs, from which we pruned noisy KG-paths, curating a high-quality subset of 25.9k. Our baseline evaluation shows an absolute scale increase by approximately 0.12 to 0.36 points for the semantic similarity score in KG-RAG over vanilla QA across multiple languages and multiple models, demonstrating the potential of KG integration. We anticipate MultiHal will foster future research towards several graph-based hallucination mitigation and fact-checking tasks.
Community
Benchmark paper for expanding current state of the art factual QA datasets with mined KG paths. Useful for factual language modeling, hallucination evaluation, KG-based knowledge updating, comparisons of knowledge updating methods (RAG vs KG-RAG)
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- ResNetVLLM-2: Addressing ResNetVLLM's Multi-Modal Hallucinations (2025)
- Generate, but Verify: Reducing Hallucination in Vision-Language Models with Retrospective Resampling (2025)
- Mitigating Hallucination in VideoLLMs via Temporal-Aware Activation Engineering (2025)
- Kaleidoscope: In-language Exams for Massively Multilingual Vision Evaluation (2025)
- Osiris: A Lightweight Open-Source Hallucination Detection System (2025)
- Grounded in Context: Retrieval-Based Method for Hallucination Detection (2025)
- Know3-RAG: A Knowledge-aware RAG Framework with Adaptive Retrieval, Generation, and Filtering (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper