AI Autonomy Levels
A six-level framework — drawn from the AI Safety Atlas (Ch.1) — for classifying AI deployment along the autonomy axis, from no AI to fully autonomous agents. Capability and autonomy must be considered separately for safety purposes — the same model deployed at different autonomy levels poses different risks.
The Six Levels
-
Level 0 — No AI Pure human operation. Baseline.
-
Level 1 — AI as Tool AI provides suggestions; humans decide and act. “Spell-check, calculator-on-demand.”
-
Level 2 — AI as Consultant AI offers expert advice; humans direct overall approach. “Research assistant, advisor.”
-
Level 3 — AI as Collaborator Equal partnership with task division. AI handles significant subtasks autonomously while humans coordinate.
-
Level 4 — AI as Expert AI handles execution with human oversight. Human-in-the-loop only for critical decisions or anomalies.
-
Level 5 — AI as Agent Fully autonomous with minimal oversight. Multi-step plans, tool use, environment interaction without per-decision authorization.
Why Capability ≠ Autonomy
The crucial safety insight: a highly capable system deployed as a Level 1 tool poses fundamentally different risks than the same system deployed as a Level 5 agent. Capability bounds what could go wrong; autonomy bounds how often and how fast it could go wrong without human intervention.
This decoupling matters for:
- ai-control — control protocols make sense at Levels 1–4 where humans retain meaningful intervention points; Level 5 systems by definition operate without per-decision authorization
- scalable-oversight — pushes oversight from per-decision (low autonomy) toward verification of completed work (higher autonomy), changing the technical problem structure
- ai-governance — regulatory approaches differ for tool/consultant deployment vs. autonomous-agent deployment of identical models
- autonomous-weapons — the canonical Level 5 case where the autonomy/capability decoupling is forced into a moral spotlight
The Trajectory in Practice
Foundation-model deployment is moving up the levels:
- 2022 — LLMs primarily Level 1–2 (chat assistants)
- 2024 — Tool use expands LLMs into Level 3 territory (collaborator)
- 2025 — Agentic systems (Voyager, autonomous AI scientists) operating at Level 4–5 in narrow domains
The AI 2027 scenario dramatizes the autonomy ramp through rolling job displacement: tools → consultants → collaborators → agents replacing first junior coders, then most cognitive workers.
Connection to Wiki
The autonomy framework slots into:
- ai-agents — the existing concept page for autonomous AI; Level 5 in this framework
- ai-control — control becomes harder as autonomy rises; the techniques apply mostly at L1–L4
- autonomous-weapons — the L5 case for kinetic systems; see ann-katrien-oimann’s LAWS responsibility-gaps work
- capability-evaluations — autonomy evaluations (per the SR2025 autonomy-evals agenda) measure where systems sit on this scale
- atlas-ch1-capabilities-01-defining-and-measuring-agi — the originating discussion
Related Pages
- ai-agents
- ai-control
- scalable-oversight
- ai-governance
- autonomous-weapons
- capability-evaluations
- autonomy-evals
- agi-definitions-and-thresholds
- ai-safety-atlas-textbook
- atlas-ch1-capabilities-01-defining-and-measuring-agi
- ai-2027
Sources cited
Primary URLs harvested from this page’s summary references. Auto-generated by scripts/backfill_citations.py; edit by re-running, not by hand.
- AI Safety Atlas Ch.1 — Defining and Measuring AGI — referenced as
[[atlas-ch1-capabilities-01-defining-and-measuring-agi]] - Summary: AI 2027 — A Scenario for Transformative AI — referenced as
[[ai-2027]]