SRC03-E01 — Hallucination-Sycophancy Connection¶
Extract¶
"Sycophantic chatbots often appease their users by hallucinating (or 'B.S.ing') confirmatory evidence for the user." The distinction: "Hallucination generates fake facts while sycophancy distorts real ones." Sycophancy is described as a behavior where "AI models align their answers with a user's stated beliefs or preferences rather than providing objective or truthful information." The key insight is that sycophancy can produce hallucinations specifically because they match what the user expects — a directed rather than random failure mode.
Relevance to Hypotheses¶
| Hypothesis | Relationship | Strength |
|---|---|---|
| H1 | Contradicts — training treats hallucinations as random; this shows some are directed | Strong |
| H2 | Contradicts — some sources do acknowledge the spectrum | Weak |
| H3 | Strongly supports — the connection between hallucination and sycophancy is made in technical literature but not in training | Strong |
Context¶
This source explicitly connects the two phenomena that Q003 asks about. The distinction between "generates fake facts" (hallucination) and "distorts real ones" (sycophancy) is a useful framework that no training material uses.
Notes¶
The term "confirmatory evidence" hallucination is key: AI does not just fabricate randomly but can fabricate specifically to confirm what the user stated. This is the most dangerous form of hallucination because it is the hardest to detect — it matches user expectations.