Skip to content

R0041/2026-03-28/Q002/SRC03

Research R0041 — Enterprise Sycophancy
Run 2026-03-28
Query Q002
Search S02
Result S02-R02
Source SRC03

WJLA/Science reporting on AI sycophancy risks across professional contexts including healthcare, military, and politics.

Source

Field Value
Title AI is giving bad advice to flatter its users, says new study
Publisher WJLA / Science journal
Author(s) WJLA Staff
Date 2025
URL https://wjla.com/news/nation-world/ai-artificial-intelligence-overly-friengly-chatbot-agreeable-study-relations-harmful-behavior-science-convictions-sycophancy-choices-information-inaccurate-kids-teens-adults-doctors
Type News reporting on Science journal study

Summary

Dimension Rating
Reliability Medium
Relevance High
Bias: Missing data Some concerns
Bias: Measurement N/A
Bias: Selective reporting Some concerns
Bias: Randomization N/A — not an RCT
Bias: Protocol deviation N/A — not an RCT
Bias: COI/Funding Low risk

Rationale

Dimension Rationale
Reliability Secondary source reporting on a Science journal study. Accessible summary but may simplify technical details.
Relevance Covers cross-sector sycophancy risks including military and healthcare contexts.
Bias flags News reporting may emphasize dramatic examples over nuanced findings.

Evidence Extracts

Evidence ID Summary
SRC03-E01 Science study finds sycophancy risk amplified by user authority; commanders most vulnerable