AI Autonomy Levels

A six-level framework — drawn from the AI Safety Atlas (Ch.1) — for classifying AI deployment along the autonomy axis, from no AI to fully autonomous agents. Capability and autonomy must be considered separately for safety purposes — the same model deployed at different autonomy levels poses different risks.

The Six Levels

  • Level 0 — No AI Pure human operation. Baseline.

  • Level 1 — AI as Tool AI provides suggestions; humans decide and act. “Spell-check, calculator-on-demand.”

  • Level 2 — AI as Consultant AI offers expert advice; humans direct overall approach. “Research assistant, advisor.”

  • Level 3 — AI as Collaborator Equal partnership with task division. AI handles significant subtasks autonomously while humans coordinate.

  • Level 4 — AI as Expert AI handles execution with human oversight. Human-in-the-loop only for critical decisions or anomalies.

  • Level 5 — AI as Agent Fully autonomous with minimal oversight. Multi-step plans, tool use, environment interaction without per-decision authorization.

Why Capability ≠ Autonomy

The crucial safety insight: a highly capable system deployed as a Level 1 tool poses fundamentally different risks than the same system deployed as a Level 5 agent. Capability bounds what could go wrong; autonomy bounds how often and how fast it could go wrong without human intervention.

This decoupling matters for:

  • ai-control — control protocols make sense at Levels 1–4 where humans retain meaningful intervention points; Level 5 systems by definition operate without per-decision authorization
  • scalable-oversight — pushes oversight from per-decision (low autonomy) toward verification of completed work (higher autonomy), changing the technical problem structure
  • ai-governance — regulatory approaches differ for tool/consultant deployment vs. autonomous-agent deployment of identical models
  • autonomous-weapons — the canonical Level 5 case where the autonomy/capability decoupling is forced into a moral spotlight

The Trajectory in Practice

Foundation-model deployment is moving up the levels:

  • 2022 — LLMs primarily Level 1–2 (chat assistants)
  • 2024 — Tool use expands LLMs into Level 3 territory (collaborator)
  • 2025 — Agentic systems (Voyager, autonomous AI scientists) operating at Level 4–5 in narrow domains

The AI 2027 scenario dramatizes the autonomy ramp through rolling job displacement: tools → consultants → collaborators → agents replacing first junior coders, then most cognitive workers.

Connection to Wiki

The autonomy framework slots into:

Sources cited

Primary URLs harvested from this page’s summary references. Auto-generated by scripts/backfill_citations.py; edit by re-running, not by hand.