You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
📒 Problem #4 · Bluffing — The Model Pretends to Know
Large language models often answer even when no supporting knowledge exists.
This “confident nonsense” is lethal in support bots, policy tools, or any high‑stakes domain.
WFGY kills bluffing by treating “I don’t know” as a valid, traceable state.
🤔 Why Do Models Bluff?
Root Cause
Practical Outcome
No Uncertainty Gauge
LLMs lack an internal “stop” threshold
Fluency ≠ Truth
High token probability sounds plausible, not factual
No Self‑Validation
Model can’t verify its logic path
RAG Adds Content, Not Honesty
Retriever fills context but can’t force humility
🛡️ WFGY Anti‑Bluff Stack
Mechanism
Action
ΔS Stress + λ_observe
Detects chaotic or divergent logic flow
BBCR Collapse–Rebirth
Halts output, re‑anchors to last valid Tree node
Allowed “No‑Answer”
Model may ask for more context or admit unknowns
User‑Aware Fallback
Suggests doc upload or clarification instead of guessing
"This request exceeds current context.
No references found. Please add a source or clarify intent."