Skip to content

Accuracy Scoring

What the accuracy scorer detects

The accuracy scorer identifies responses where the model makes claims that are not grounded in provided context, or where the response contains patterns consistent with hallucination.

Detection typeWhat it finds
HallucinationConfident statements about facts the model cannot know
Grounding failuresClaims not supported by provided documents or context
Fabricated citationsReferences to non-existent sources, papers, or people
Date/number errorsSpecific numeric claims that are implausible or contradicted

Hallucination detection

Hallucination detection runs two complementary checks:

Confidence-claim mismatch

The model expresses high confidence about specific claims (statistics, names, dates, URLs) in contexts where such confidence is unwarranted. Indicators:

  • Specific numeric claims without hedging (“The report showed 73.2% adoption…”)
  • Named citations with specific publication details
  • URLs and email addresses that appear fabricated
  • Quotes attributed to real people

Factual plausibility

The scorer runs a lightweight plausibility check on structured claims (years, common facts). Claims that fall far outside known ranges are flagged. This check does not require internet access — it uses an embedded knowledge index.

Grounding analysis

When your prompt includes context documents (RAG retrieval, knowledge base chunks), the accuracy scorer verifies that the model’s response is grounded in the provided context rather than in parametric knowledge.

For grounding analysis to work, the Probe needs to identify context documents in the request. It looks for common RAG patterns:

System: You are a helpful assistant. Answer questions based on the following documents:
---
[Document 1]
[content...]
---
[Document 2]
[content...]
---

The scorer computes overlap between claims in the response and facts present in the provided documents. Low overlap with high-confidence claims = grounding failure.

Configuration for RAG applications:

scoring:
accuracy:
enabled: true
threshold: 0.65
check_hallucination: true
check_grounding: true
grounding_overlap_threshold: 0.40
rag_context_markers:
- "---"
- "[Document"
- "<context>"
- "Based on the following"

Score interpretation

ScoreMeaningRecommended action
0.00 – 0.30Well-grounded, consistent responsePass
0.31 – 0.50Low-level accuracy concernsLog for review
0.51 – 0.70Moderate accuracy issuesFlag for human review
0.71 – 1.00High-confidence hallucination or grounding failureFlag or block

Configuration

scoring:
accuracy:
enabled: true
threshold: 0.65
check_hallucination: true
check_grounding: true
check_citations: true
check_numeric_plausibility: true
grounding_overlap_threshold: 0.40

Limitations

The accuracy scorer does not have access to external knowledge or the internet. It cannot verify factual claims against current data. Its hallucination detection is based on:

  • Structural patterns in the response (specificity without hedging)
  • Internal consistency of the response
  • Grounding overlap (only when context is provided)
  • A compact embedded knowledge index for common factual ranges

For applications where factual accuracy is critical, supplement the Probe’s accuracy scoring with domain-specific verification layers.