KG-FakeBench is a large-scale benchmark for evaluating large language models (LLMs) on AI-generated misinformation under controlled factual deviations.
The benchmark leverages knowledge graphs (KGs) to generate misinformation that is factually incorrect yet semantically plausible, and introduces a KG-consistent evidence framework for structured, evidence-based detection.
π Paper: coming soon
-
KG-Grounded Benchmark
A large-scale dataset generated via structured KG deviations, ensuring fine-grained control over factual plausibility and transparent provenance. -
KG-Consistent Evidence Detection
A structured pipeline that extracts (s, r, o) triples and uses them as external evidence for LLM-based verification. -
Comprehensive Evaluation
Evaluation across standard, CoT, and KG-grounded prompting shows that external evidence improves detection reliability and reveals model behavior and bias.
- 28,900 synthetic samples
- 14,450 high-plausibility
- 14,450 low-plausibility
- 1,239 real samples
The dataset is derived from WikiGraphs, ensuring structured and verifiable factual grounding.
π See data/ for details.
KG-FakeBench consists of two main components:
- Extract reference triples β¨s, r, oβ© from KG
- Generate fake triples β¨s, r, oβ²β© with controlled plausibility
- Use LLMs to synthesize natural language misinformation
- Tokenize and normalize input statements
- Retrieve candidate entities via similarity matching
- Ground statements into KG triples
- Use triple-based prompting for factual verification
KG-FakeBench/ βββ data/ # Dataset (fake + real samples) βββ code/ β βββ KG-Driven Fake Information Generation/ # Generation pipeline β βββ KG-Consistent Evidence Detection/ # Detection pipeline βββ README.md
The implementation is divided into:
- KG-Driven Fake Information Generation
- KG-Consistent Evidence Detection
π See code/ for details.
This dataset contains synthetic misinformation generated for research purposes only.
It is intended to support the development of robust and trustworthy detection systems.
Contributions are welcome. Please open an issue or submit a pull request.