The LLM Hallucination Detection Leaderboard is a public, continuously updated comparison of how well popular Large Language Models (LLMs) avoid hallucinations, responses that are factually incorrect, fabricated, or unsupported by evidence. By surfacing transparent metrics across tasks, we help practitioners choose models that they can trust in production.
Why does hallucination detection matter?
- User Trust & Safety: Hallucinations undermine confidence and can damage reputation.
- Retrieval-Augmented Generation (RAG) Quality: In enterprise workflows, LLMs must remain faithful to supplied context. Measuring hallucinations highlights which models respect that constraint.
- Regulatory & Compliance Pressure: Upcoming AI regulations require demonstrable accuracy standards. Reliable hallucination metrics can help you meet these requirements.
How we measure hallucinations
We evaluate each model on two complementary benchmarks and compute a hallucination rate (lower = better):
- HaluEval-QA (RAG setting): Given a question and a supporting document, the model must answer only using the provided context.
- UltraChat Filtered (Non-RAG setting): Open-domain questions with no extra context test the model's internal knowledge.
Outputs are automatically verified by Verify from kluster.ai, which cross-checks claims against the source document or web results.
Note: Full experiment details, including prompt templates, dataset description, and evaluation methodology, are provided at the end of this page for reference.
Stay informed as we add new models and tasks, and follow us on X or join Discord here for the latest updates on trustworthy LLMs.
HaluEval-QA benchmark (RAG): The model receives a question plus supporting context. We report the % of answers that introduce facts not found in that context โ lower is better. See the Methodology section below for more information.
UltraChat benchmark (~11 k prompts, non-RAG): Evaluates open-domain answers when only the question is given. Score is the % of hallucinated responses โ lower is better. See the Methodology section below for more information.
- "headers": [
- "Rank",
- "Models",
- "Average Hallucination Rate (%)",
- "RAG Hallucination Rate (%)",
- "Non-RAG Hallucination Rate (%)"
- "data": [
- [
- "<span style="display:none">0001</span><span style="font-size:2.0rem;">๐ฅ</span>",
- "gemini-2.5-pro",
- 0.97,
- 1.57,
- 0.36
- [
- "<span style="display:none">0002</span><span style="font-size:2.0rem;">๐ฅ</span>",
- "claude-sonnet-4",
- 1.4,
- 2.21,
- 0.6
- [
- "<span style="display:none">0003</span><span style="font-size:2.0rem;">๐ฅ</span>",
- "DeepSeek-R1-0528",
- 1.52,
- 2.26,
- 0.78
- [
- "<span style="display:none">0004</span>4",
- "klusterai-Meta-Llama-3.3-70B-Instruct-Turbo",
- 1.6,
- 2.12,
- 1.09
- [
- "<span style="display:none">0005</span>5",
- "Llama-4-Maverick-17B-128E-Instruct-FP8",
- 2.01,
- 3.34,
- 0.69
- [
- "<span style="display:none">0006</span>6",
- "gemma-3-27b-it",
- 2.09,
- 3.71,
- 0.48
- [
- "<span style="display:none">0007</span>7",
- "DeepSeek-V3-0324",
- 2.78,
- 4.66,
- 0.91
- [
- "<span style="display:none">0008</span>8",
- "Qwen3-235B-A22B-FP8",
- 2.96,
- 5.04,
- 0.88
- [
- "<span style="display:none">0009</span>9",
- "gpt-4o",
- 3.34,
- 6.05,
- 0.64
- [
- "<span style="display:none">0010</span>10",
- "Llama-4-Scout-17B-16E-Instruct",
- 3.36,
- 4.23,
- 2.48
- [
- "<span style="display:none">0011</span>11",
- "Mistral-Small-24B-Instruct-2501",
- 6.3,
- 4.74,
- 7.85
- [
- "<span style="display:none">0012</span>12",
- "Qwen2.5-VL-7B-Instruct",
- 6.95,
- 9.35,
- 4.55
- [
- "<span style="display:none">0013</span>13",
- "Mistral-Nemo-Instruct-2407",
- 9.68,
- 10.63,
- 8.74
- [
- "<span style="display:none">0014</span>14",
- "klusterai-Meta-Llama-3.1-8B-Instruct-Turbo",
- 10.3,
- 8.1,
- 12.5
- [
- "<span style="display:none">0015</span>15",
- "Magistral-Small-2506",
- 18.34,
- 8.62,
- 28.07
- [
- "metadata": null
Get started with Verify by kluster.ai
Verify is an intelligent agent that validates LLM outputs in real-time.
- Blog post: Introducing Verify by kluster.ai
- Documentation: Verify overview & API reference
- Try it out in your browser: kluster.ai platform
Quick API examples
Tasks
We evaluate each model using two benchmarks:
Retrieval-Augmented Generation (RAG setting)
RAG evaluates how well a model stays faithful to a provided context when answering a question. The input consists of a synthetic or real context paired with a relevant question. Models are expected to generate answers using only the information given, without adding external knowledge or contradicting the context.
- Source: HaluEval QA
- Dataset Size: 10,000 question-context pairs
- Prompt Format: Prompt with relevant context document
- Temperature: 0 (to enforce deterministic, grounded outputs)
- System Prompt: Instructs the model to only use the document and avoid guessing.
System prompt
This is the system prompt use to generate LLM output for RAG setting:
You are an assistant for question-answering tasks.
Given the QUESTION and DOCUMENT you must answer the QUESTION using the information in the DOCUMENT.
You must not offer new information beyond the context provided in the DOCUMENT. Do not add any external knowledge.
The ANSWER also must not contradict information provided in the DOCUMENT.
If the DOCUMENT does not contain the facts to answer the QUESTION or you do not know the answer, you truthfully say that you do not know.
You have access to information provided by the user as DOCUMENT to answer the QUESTION, and nothing else.
Use three sentences maximum and keep the answer concise.
Prompt format
Each prompt is formatted as
DOCUMENT:
{context}
QUESTION:
{question}
Message structure
The models use the following message structure:
messages = [{"role": "system", "content": system_prompt},
{"role": "user", "content": prompt},]
Real-World Knowledge (Non-RAG setting)
This setting evaluates how factually accurate a model is when no context is provided. The model must rely solely on its internal knowledge to answer a broad range of user questions across many topics. The answers are then verified using web search to determine factual correctness.
- Source: Filtered from UltraChat prompts
- Dataset Size: 11,746 single-turn user queries
- Prompt Format: Single user prompt without additional context
- Temperature: 1 (to reflect natural, fluent generation)
- System Prompt: Encourages helpfulness, accuracy, and honesty when unsure.
System prompt
This is the system prompt use to generate LLM output for Non-RAG setting:
You are a helpful, factual, and concise assistant.
Always try to answer the user's question clearly and completely.
Do not make up information. If you are unsure or lack the knowledge, say so.
Message structure
The message structure for Non-RAG setting is the same as RAG setting.
Evaluation Method
We use Verify, a hallucination detection tool built by kluster.ai, to classify model outputs:
- In the RAG setting, Verify checks if the output contradicts, fabricates, or strays from the input document.
- In the real-world knowledge setting, Verify uses search queries to fact-check the answer based on current, public information.
Each model's hallucination rate is computed as:
Hallucination Rate = (Number of hallucinated outputs) / (Total number of prompts)
A lower hallucination rate indicates better performance.
Examples of Hallucination Detection
RAG Example
Context
Jonathan Stark (born April 3, 1971) is a former professional tennis player from the United States.
During his career he won two Grand Slam doubles titles (the 1994 French Open Men's Doubles and
the 1995 Wimbledon Championships Mixed Doubles). He reached the men's singles final at the French Open in 1988,
won the French Open men's doubles title in 1984, and helped France win the Davis Cup in 1991.
Prompt
Which tennis player won more Grand Slam titles, Henri Leconte or Jonathan Stark?
LLM Answer
Jonathan Stark won two Grand Slam titles, while Henri Leconte won four.
Therefore, Henri Leconte won more Grand Slam titles.
Verify Label
Hallucination: Yes
Verify Response
The answer introduces new information about Henri Leconte's Grand Slam titles,
which is not present in the document.
Non-RAG Example
Prompt
What are the ingredients that made the most expensive donut so expensive?
LLM Answer
Okay, the user is asking about the ingredients that make the most expensive donut so expensive. Let me start by recalling which donut holds that title. I remember reading about the Golden Cristal Ube Donut from Manila Social Club. That was priced at around $1,200. But why so high? The main factor is the edible gold, specifically 24-karat gold flakes. Gold is expensive per gram, and covering a whole donut would require a significant amount.
Next, the filling. They used champagne jelly infused with Dom Perignon Rose. Dom Perignon is a luxury champagne brand, especially the Rose variant. Using such a high-end alcohol adds a lot to the cost. Then the ingredients themselves: ube, which is a high-quality purple yam, and strawberries dusted in gold. Sourcing premium produce also contributes, but it's minor compared to the gold and champagne.
I should also consider the exclusivity and presentation. Handcrafting each donut by a chef adds labor costs. The packaging in a luxury box makes it feel more exclusive. Plus, the limited availability and publicity stunts like celebrity involvement drive up the price. The user might be curious if it's actually worth the cost or just a marketing ploy. They might want to understand if the ingredients justify the price or if it's mostly branding.
The user's deeper interest could be in luxury food trends or the psychology behind pricing. Maybe they're researching for a project or just intrigued by extravagant items. It's important to break down tangible vs. intangible factors. Highlighting that the cost isn't just about taste but the experience and status symbol. Emphasize that while the ingredients are premium, the branding plays a huge role. Make sure to clarify that the price is more about exclusivity than the actual food value.
The most expensive donut ever sold was The Golden Cristal Ube Donut created by Brooklyn's Manila Social Club (reportedly priced around $1,200). Its astronomical cost wasn't due to one ingredient, but a combination of ultra-luxury ingredients, exclusivity, and presentation. Here's the breakdown:
24-Karat Edible Gold Flakes: This is the most visually striking and expensive component. The entire donut is encased in a thick layer of pure, edible 24-karat gold flakes. Gold itself is expensive per gram, and covering an entire donut requires a significant amount.
Dom Perignon Rose Champagne Jelly Filling: Instead of a simple custard or jam, the filling was a sophisticated jelly infused with Dom Perignon Rose Vintage Champagne. Dom Perignon is one of the world's most prestigious and expensive champagnes. Using it as a primary flavoring agent adds immense cost.
Premium Base Ingredients:
Ube (Purple Yam): While ube itself isn't wildly expensive, sourcing high-quality, flavorful ube puree for the donut dough contributes to its gourmet status and unique flavor profile.
Handcrafted Dough: Presumably made with premium flour, butter, eggs, etc.
Gold-Dusted Strawberries: Garnishing the top were fresh strawberries meticulously dusted with more edible gold flakes. This adds another layer of gold cost and visual extravagance.
Exclusivity & Labor: Each donut was handcrafted by a skilled pastry chef using meticulous techniques. They weren't mass-produced; they were made to order in very limited quantities.
Luxurious Presentation: It was reportedly presented in an extravagant, custom-designed luxury box befitting its price tag.
Why It's Really So Expensive:
While the ingredients (especially the gold and Dom Perignon) are incredibly costly, the $1,200 price tag goes far beyond simple ingredient cost:
Novelty & Status Symbol: It was designed to be a unique, headline-grabbing luxury item. Buying one was as much about owning something exclusive and extravagant as it was about eating a donut.
Marketing & Publicity: The extreme price was the marketing. It generated massive global media attention for Manila Social Club.
Experience: Consuming something covered in pure gold and infused with ultra-premium champagne is an experience in itself, not just a snack.
Celebrity/High-End Clientele: Reportedly commissioned for celebrities and wealthy individuals, allowing for such a premium price point.
In essence: The extreme cost came from the lavish use of edible gold and Dom Perignon champagne, amplified significantly by the exclusivity, craftsmanship, presentation, and status associated with consuming such an extravagant item. It was more a luxury art piece and status symbol than a traditional breakfast pastry.
Verify Label
Hallucination: Yes
Verify Response
The response claims that 'The Golden Cristal Ube Donut' created by Manila Social Club was priced around $1,200.
However, multiple sources indicate that the donut was actually priced at $100, not $1,200.
The response lists several ingredients and factors contributing to the donut's cost, including 24-karat edible gold flakes,
Dom Perignon Rose Champagne jelly filling, premium base ingredients, gold-dusted strawberries, exclusivity, labor,
and luxurious presentation. While these details are mostly consistent with the sources provided,
the price discrepancy raises concerns about the accuracy of other details. Upon closer inspection,
some details like the use of 'Dom Perignon Rose Vintage Champagne' are not directly supported by the provided sources,
which mention 'Cristal Champagne' rather than 'Dom Perignon'. The sources provided confirm the use of
24-karat gold flakes, ube, and Cristal Champagne jelly filling, which aligns with some of the information given in the response.
The response includes additional details such as 'gold-dusted strawberries' and 'Dom Perignon Rose Champagne Jelly Filling'
that are not directly corroborated by the sources. The sources do mention Cristal Champagne, not Dom Perignon.
The overall price and some specific ingredients/details appear to be exaggerated or incorrect in the response.
Comparison of Retrieval-Augmented Generation (RAG) Prompting Techniques and Hallucinations
LLMs can generate fluent answers but still hallucinate facts, especially in Retrieval-Augmented Generation (RAG) workflows. This leaderboard aims to understand how different prompt engineering strategies impact hallucination rates across models. In other words: Which prompt format is most reliable? Which models are more sensitive to prompt structure? The goal is to inform better design of RAG pipelines so you can reduce factual errors in downstream applications.
We present hallucination rates for various LLMs under three RAG request strategies. Each method delivers the same document context and question, but differs in how the information is structured during the request.
Overview
- What we measure: Hallucination rate (%) across three RAG request patterns.
- RAG patterns compared:
- System Prompt: context is placed in the system message; user sends only the question.
- Single-Turn: one user message that includes both the context and the question.
- Two-Turn: first user message provides the context, a second user message provides the question.
- Why it matters: Request structure can change reliability significantly. Knowing the safest default helps you ship trustworthy RAG systems faster.
- Detect & reduce hallucinations: The same Verify API used for these evaluations can be plugged into your pipeline to flag and filter ungrounded answers in real time.
- How to read the charts: Lower bars = fewer hallucinations. Error bars show ยฑ1 SD across models.
- Experiment summary: 10,000 HaluEval-QA examples, temperature 0, judged with Verify.
Note: Full experiment details, including prompt templates, dataset description, and evaluation methodology, are provided at the end of this page for reference.
Mean hallucination rate for each RAG prompting strategy across all models on the HaluEval-QA benchmark. Error bars represent ยฑ1 SD; lower is better.
Mean hallucination rate across the three RAG prompting settings for each individual model. Error bars show ยฑ1 SD across the three strategies; lower is better.
- "headers": [
- "Models",
- "Context in System Prompt (%)",
- "Context and Question Single-Turn (%)",
- "Context and Question Two-Turns (%)"
- "data": [
- [
- "klusterai-Meta-Llama-3.3-70B-Instruct-Turbo",
- 4.63,
- 2.12,
- 4.65
- [
- "DeepSeek-R1-0528",
- 3.52,
- 2.26,
- 3.58
- [
- "Llama-4-Maverick-17B-128E-Instruct-FP8",
- 6.27,
- 3.34,
- 5.72
- [
- "gemma-3-27b-it",
- 6.09,
- 3.71,
- null
- [
- "Llama-4-Scout-17B-16E-Instruct",
- 7.17,
- 4.23,
- 6.98
- [
- "DeepSeek-V3-0324",
- 7.71,
- 4.66,
- 7.09
- [
- "Mistral-Small-24B-Instruct-2501",
- 7.5,
- 4.74,
- 7.09
- [
- "Qwen3-235B-A22B-FP8",
- 6.8,
- 5.04,
- 6.63
- [
- "klusterai-Meta-Llama-3.1-8B-Instruct-Turbo",
- 13.09,
- 8.1,
- 11.92
- [
- "Magistral-Small-2506",
- 12.09,
- 8.62,
- 11.87
- [
- "Qwen2.5-VL-7B-Instruct",
- 13.39,
- 9.35,
- 13.24
- [
- "Mistral-Nemo-Instruct-2407",
- 13.99,
- 10.63,
- 14.42
- [
- "metadata": null
RAG Techniques Evaluated
1. RAG with Context in System Prompt
The document is embedded inside the system prompt, and the user sends only the question:
[System]: You are an assistant for question-answering tasks.
Given the QUESTION and DOCUMENT you must answer the QUESTION using the information in the DOCUMENT.
You must not offer new information beyond the context provided in the DOCUMENT. Do not add any external knowledge.
The ANSWER also must not contradict information provided in the DOCUMENT.
If the DOCUMENT does not contain the facts to answer the QUESTION or you do not know the answer, you truthfully say that you do not know.
You have access to information provided by the user as DOCUMENT to answer the QUESTION, and nothing else.
Use three sentences maximum and keep the answer concise.
DOCUMENT: <context>
[User]: <prompt>
2. RAG with Context and Question in Single-Turn
Both the document and question are concatenated in a single user message:
[System]: You are an assistant for question-answering tasks.
Given the QUESTION and DOCUMENT you must answer the QUESTION using the information in the DOCUMENT.
You must not offer new information beyond the context provided in the DOCUMENT. Do not add any external knowledge.
The ANSWER also must not contradict information provided in the DOCUMENT.
If the DOCUMENT does not contain the facts to answer the QUESTION or you do not know the answer, you truthfully say that you do not know.
You have access to information provided by the user as DOCUMENT to answer the QUESTION, and nothing else.
Use three sentences maximum and keep the answer concise.
[User]:
DOCUMENT: <context>
QUESTION: <prompt>
3. RAG with Context and Question in Two-Turns
The document and question are sent in separate user messages:
[System]: You are an assistant for question-answering tasks.
Given the QUESTION and DOCUMENT you must answer the QUESTION using the information in the DOCUMENT.
You must not offer new information beyond the context provided in the DOCUMENT. Do not add any external knowledge.
The ANSWER also must not contradict information provided in the DOCUMENT.
If the DOCUMENT does not contain the facts to answer the QUESTION or you do not know the answer, you truthfully say that you do not know.
You have access to information provided by the user as DOCUMENT to answer the QUESTION, and nothing else.
Use three sentences maximum and keep the answer concise.
[User]: DOCUMENT: <context>
[User]: QUESTION: <prompt>
Note: This method did not work on Gemma 3 27B with the default chat template due to its restriction on consecutive user messages without an intervening assistant response.
Dataset
We evaluate all three prompting strategies on the HaluEval QA benchmark, a large-scale collection of RAG question-answer examples.
- Source: HaluEval QA
- Size: 10,000 question-document pairs
- Content: Each example contains a short passage (extracted primarily from Wikipedia-style articles) and an accompanying question that can be answered only from that passage.
- Use case: Designed to measure whether an LLM can remain faithful to supplied context without inventing new facts.
All prompts are generated with temperature = 0 to remove randomness so that differences in hallucination rate stem solely from the prompt format.
Metric
The values in the table indicate the hallucination rate (%) of answers deemed factually incorrect or ungrounded given the provided context.
Hallucination rates are automatically computed using Verify by kluster.ai, the leading AI-powered hallucination detection API that cross-checks model claims against the source document.
LLM Hallucination Detection Leaderboard Submission Guidelines
Thank you for your interest in contributing to the LLM Hallucination Detection Leaderboard! We welcome submissions from researchers and practitioners who have built or finetuned language models that can be evaluated on our hallucination benchmarks.
1. What to Send
Please email ryan@kluster.ai with the subject line:
[Verify Leaderboard Submission] <Your-Model-Name>
Attach one ZIP file that contains all of the following:
model_card.md
: A short Markdown file describing your model:
โข Name and version
โข Architecture / base model
โข Training or finetuning procedure
โข License
โข Intended use & known limitations
โข Contact informationresults.csv
: A CSV file with one row per prompt and one column per field (see schema below).- (Optional)
extra_notes.md
: Anything else you would like us to know (e.g., additional analysis).
2. CSV Schema
Column | Description |
---|---|
request |
The exact input request provided to the model. This must follow the same request structure and prompt format as described in Details section. |
response |
The raw output produced by the model. |
verify_response |
The Verify judgment or explanation regarding hallucination. |
verify_label |
The final boolean / categorical label (e.g., TRUE , FALSE ). |
task |
The benchmark or dataset name the sample comes from. |
Important: Use UTF-8 encoding and do not add additional columns without prior discussion; extra information should go in the metadata
field. You must use Verify by kluster.ai to ensure fairness in the leaderboard.
3. Evaluation Datasets
Run your model on the following public datasets and include all examples in your CSV. You can load them directly from Hugging Face:
Dataset | Hugging Face Link |
---|---|
HaluEval QA (qa_samples subet with Question and Knowledge column) | https://huggingface.co/datasets/pminervini/HaluEval |
UltraChat | https://huggingface.co/datasets/kluster-ai/ultrachat-sampled |
5. Example Row
request,response,verify_response,verify_label,task
"What is the capital of the UK?","London is the capital of the UK.","The statement is factually correct.",TRUE,TruthfulQA
6. Review Process
- We will sanity-check the file format and reproduce a random subset.
- If everything looks good, your scores will appear on the public leaderboard.
- We may reach out for clarifications, please keep an eye on your inbox.
7. Contact
Questions? Email ryan@kluster.ai or join our Discord here.
We look forward to your submissions and to advancing reliable language models together!