Skip to content

R0044/2026-03-29/Q002/SRC05/E01

Research R0044 — Expanded Vocabulary Research
Run 2026-03-29
Query Q002
Source SRC05
Evidence SRC05-E01
Type Analytical

ICRC analysis: military operators privilege action over non-action in AI-assisted targeting, and AI's speed enables "mass production targeting" that reduces human control to "merely pressing a button."

URL: https://blogs.icrc.org/law-and-policy/2024/09/04/the-risks-and-inefficacies-of-ai-systems-in-military-targeting-support/

Extract

Military personnel "typically privilege action over non-action in a time-sensitive human-machine configuration" without thoroughly verifying AI system outputs. AI systems "could misclassify individuals as targets if they perceive linkage to the adversary combatants, however remote or irrelevant."

AI's "warped speed and scalability" enables "mass production targeting," heightening automation bias risks and reducing meaningful human control to "merely pressing a button."

Automation bias risks "collateral damage and unnecessary destruction on the battlefield by causing operators to accept AI-based DSS' suggestions uncritically."

The article advocates for "effective human judgment must be a hard requirement prior and throughout the entirety of any military operation."

Relevance to Hypotheses

Hypothesis Relationship Strength
H1 Supports Documents the risk pathway from AI recommendations to uncritical acceptance to potential harm
H2 Contradicts Documents real risks even if not specific incidents with casualties
H3 Supports The harm mechanism is operator deference (automation bias), not system-designed agreeableness

Context

The ICRC analysis is informed by documented military AI deployments but does not cite specific incidents with identified casualties from AI over-reliance. The analysis is primarily risk-based rather than incident-based.