-
Notifications
You must be signed in to change notification settings - Fork 0
feat: cross-provider uncertainty check for hallucination detection at approval gates #701
Description
Context
MIT research proposes a combined epistemic + aleatoric uncertainty metric (TU) that cross-references responses across multiple LLM providers to flag hallucinations more reliably than self-consistency alone.
Why This Matters
SynthOrg already abstracts multiple LLM providers via BaseCompletionProvider. A lightweight cross-provider consistency check at approval gates is architecturally feasible. This augments progressive trust and the LLM fallback evaluator with a concrete hallucination signal.
Action Items
- Design cross-provider consistency check: same prompt to 2+ providers, measure agreement
- Integrate as optional signal in approval gate logic (high disagreement = flag for human review)
- Evaluate cost trade-off: extra provider call vs hallucination prevention value
- Cross-reference with Rozum AI verification pattern (76.2% unsupported claims flagged)
References
- MIT research
- Rozum AI -- same multi-model verification concept
Additional Research (2026-03-26)
Multi-Model Verification Data
Source: Rozum AI (2026-03-22)
Concrete validation of the cross-provider approach: Rozum AI's multi-model verification layer flagged unsupported claims in 76.2% of outputs. This demonstrates that cross-model disagreement is a strong signal for hallucination detection.
Two-Stage Classifier (related)
See #847 for a complementary approach: Anthropic's two-stage safety classifier with information stripping. The cross-provider uncertainty check (#701) and the two-stage classifier (#847) can work in tandem -- #701 for content verification, #847 for action authorization.