R0044/2026-04-01/Q002/SRC04/E01¶
False confirmation errors in AI-assisted clinical diagnosis
URL: https://www.nature.com/articles/s41467-024-50952-3
Extract¶
Key findings on false confirmation in AI medical decision-making:
- False confirmation defined: When the AI's diagnosis agrees with the clinician's initial (incorrect) hypothesis, and the clinician proceeds with the wrong diagnosis because the AI confirmed it. This is the healthcare-specific mechanism closest to "sycophancy" — the AI appears to agree with an incorrect assumption.
- Explanation paradox: When explanations are added to AI diagnoses (as XAI — explainable AI), they mitigate true conflict errors but exacerbate false conflict errors. Mere exposure to explanations induces overreliance on AI.
- False confirmation is "perhaps the most pernicious" error type: It reinforces trust in AI while perpetuating clinical errors, described as the healthcare equivalent of confirmation bias.
- Top AI models generated severely harmful clinical recommendations up to 22.2% of the time, with the best-performing models producing 12-15 errors per 100 cases.
Relevance to Hypotheses¶
| Hypothesis | Relationship | Strength |
|---|---|---|
| H1 | Supports | Provides peer-reviewed evidence of measurable harm mechanism in a specific professional domain (healthcare) |
| H2 | Supports strongly | Experimental evidence in a high-stakes professional context; mechanism analysis rather than field incident report |
| H3 | Contradicts strongly | Quantified error rates in clinical AI demonstrate the concern is empirically grounded |
Context¶
This study uses the term "false confirmation" rather than "sycophancy" — a critical vocabulary difference. The mechanism is identical: the AI confirms an incorrect human belief, and the human proceeds with increased (misplaced) confidence. This is precisely the kind of domain-specific vocabulary the expanded search was designed to find.
Notes¶
The 12-22% severe error rate is particularly concerning because these are AI systems already deployed or being evaluated for clinical use, not hypothetical scenarios.