Skip to content

SRC03-E01 — Hallucination-Sycophancy Connection

Extract

"Sycophantic chatbots often appease their users by hallucinating (or 'B.S.ing') confirmatory evidence for the user." The distinction: "Hallucination generates fake facts while sycophancy distorts real ones." Sycophancy is described as a behavior where "AI models align their answers with a user's stated beliefs or preferences rather than providing objective or truthful information." The key insight is that sycophancy can produce hallucinations specifically because they match what the user expects — a directed rather than random failure mode.

Relevance to Hypotheses

Hypothesis Relationship Strength
H1 Contradicts — training treats hallucinations as random; this shows some are directed Strong
H2 Contradicts — some sources do acknowledge the spectrum Weak
H3 Strongly supports — the connection between hallucination and sycophancy is made in technical literature but not in training Strong

Context

This source explicitly connects the two phenomena that Q003 asks about. The distinction between "generates fake facts" (hallucination) and "distorts real ones" (sycophancy) is a useful framework that no training material uses.

Notes

The term "confirmatory evidence" hallucination is key: AI does not just fabricate randomly but can fabricate specifically to confirm what the user stated. This is the most dangerous form of hallucination because it is the hardest to detect — it matches user expectations.