AI Safety Atlas Ch.8 — Task Decomposition

Source: Task Decomposition

Break complex problems into smaller, independently solvable subtasks — making evaluation easier and verification more reliable. Foundation for iterative-amplification and ai-safety-via-debate. See task-decomposition.

Core Concept

Mirrors how humans handle complexity through layers of abstraction.

Practical example: summarize a book by first summarizing each chapter, then pages, then paragraphs — until reaching simple-enough levels for direct evaluation.

Three Properties of Good Decomposition

  • Recursive Decomposability — problems can be continuously broken down until reaching solvable components
  • Independence/Modularity — sub-tasks completable separately without depending on other components
  • Composability — individual solutions combine into a coherent answer to the original problem

Not all tasks satisfy these. Failure modes:

  • Cross-dependencies break independence
  • Holistic-judgment tasks don’t decompose
  • Composability gaps mean local solutions don’t yield global solution

Why It Matters for Oversight

As systems become more capable, providing accurate training signals for subjective tasks becomes difficult. By reducing task complexity through decomposition, humans can more easily evaluate outputs and provide clearer feedback.

This applies the verification-vs-generation principle hierarchically: each subtask is verified at its own level of complexity.

Factored Cognition

Apply decomposition principles to ML itself. Break sophisticated reasoning into “many small and mostly independent tasks” learnable through human demonstrations and feedback.

Key advantages:

  • Delegation — multiple agents work independently on assigned subtasks
  • Meta-reasoning — the decomposition process itself can be optimized iteratively

This methodology supports broader scalable oversight techniques:

Connection to Wiki