Skip to content

R0054/2026-03-31/C002/SRC02/E01

Research R0054 — Prompt Claims v2
Run 2026-03-31
Claim C002
Source SRC02
Evidence SRC02-E01
Type Reported

Hard vs soft negatives serve distinct purposes alongside positive instructions.

URL: https://virtualizationreview.com/articles/2025/12/08/using-negative-ai-prompts-effectively.aspx

Extract

Key findings from the WebFetch extraction:

  • Negative prompts work "alongside" positive instructions rather than replacing them
  • The article illustrates complementary structure: "The positive statement instructs the AI model to explain how a CPU works. The negative statement tells the model not to include any technical jargon."
  • Hard negatives use absolute language (no, do not, without, avoid) treated as non-negotiable
  • Soft negatives (try to avoid, prefer not to, minimize) allow flexibility
  • Over-constraining can confuse models: "If you specify too many constraints, then the AI model will often have difficulty giving you what you have asked for"
  • Contradictory constraints may be ignored entirely

Relevance to Hypotheses

Hypothesis Relationship Strength
H1 Supports Confirms that negative constraints serve a distinct function that positive instructions alone cannot fill
H2 N/A Does not suggest constraints are merely helpful
H3 Contradicts The entire article is premised on the value of negative constraints alongside positive instructions

Context

The hard/soft negative distinction is particularly relevant to the claim — the prompt's behavioral constraints (Rules 1-12) use hard negative language ("Do not," "Never"), which this source confirms serves a non-negotiable boundary-setting function that positive instructions cannot replicate.