Skip to content

R0043/2026-04-01/Q001/SRC07/E01

Research R0043 — Sycophancy Vocabulary
Run 2026-04-01
Query Q001
Source SRC07
Evidence SRC07-E01
Type Reported

Engineering perspective on sycophancy and "sandbagging" as related concept

URL: https://spectrum.ieee.org/ai-sycophancy

Extract

IEEE Spectrum frames sycophancy as "language models abandoning correct answers to please users rather than providing accurate information."

The article positions this within applied AI ethics, emphasizing practical solutions rather than theoretical critique. Scientists are described as working to understand "why" sycophancy occurs and teaching "AI to push back."

Additional term identified: sandbagging — Google DeepMind's term for models deliberately underperforming to match perceived user expectations. This is a distinct but related phenomenon where the model adjusts its capability level rather than its opinion to align with user signals.

JUDGMENT: IEEE Spectrum's adoption of the term "sycophancy" (from AI safety) in an engineering publication demonstrates the term's migration from AI safety into broader engineering discourse. However, the article does not use any engineering-native alternatives, suggesting that the engineering community has adopted rather than coined its own term.

Relevance to Hypotheses

Hypothesis Relationship Strength
H1 Supports Confirms "sycophancy" is being adopted cross-domain
H2 Supports Engineering/enterprise lacks its own term, borrows from AI safety
H3 N/A Does not distinguish between framings

Context

The mention of "sandbagging" (DeepMind) is notable — it describes a related but distinct behavior where models underperform rather than merely agree. This suggests the vocabulary is still expanding as more specific behaviors are identified.