What counts as a hallucination?
The hallucination rate measures how often a model generates information that cannot be found in its provided inputs, such as retrieved documents, user messages, or system prompts. Quotient reports an extrinsic hallucination rate. We determine whether the model’s output is externally unsupported by the context it was given.What is an Extrinsic Hallucination?
What is an Extrinsic Hallucination?
How Quotient detects hallucinations
- Segment the output into atomic claims or sentences.
- Cross-check every claim against all available context:
user_query(what the user asked)documents(retrieved evidence)message_history(prior conversation turns)instructions(system or developer guidance)
- Flag unsupported claims when no context backs them up.
Inputs that improve detection quality
- High-signal documents: include only the evidence actually retrieved for the answer.
- Conversation history: pass the full multi-turn exchange so references to earlier turns can be validated.
- Instructions: provide system prompts so the detection pass understands guardrails and policies.
Interpret hallucination results
has_hallucination: Boolean flag indicating whether we found any unsupported claims.- Highlighted spans: In the dashboard, statements are color-coded to show what lacked support.
- Tag filters: Slice hallucination rate by model, feature, or customer to prioritize remediation.
Why monitor hallucinations?
Extrinsic hallucinations are a primary failure mode in augmented AI systems. Even when retrieval succeeds, generation can drift. Tracking this metric helps teams:- Catch hallucinations early in development.
- Monitor output quality after deployment.
- Guide prompt iteration and model fine-tuning.