AI Safety Atlas Ch.2 — Appendix: Quantifying Existential Risks

Source: Appendix: Quantifying Existential Risks

Expert subjective probability estimates for AI causing existential catastrophe — the P(doom) metric and its limitations. See p-doom for the dedicated concept page.

What P(doom) Means

P(doom) = subjective probability that AI causes existentially catastrophic outcomes for humanity. Generally encompasses extinction, permanent disempowerment, or civilizational collapse. No standardized methodology — each estimate reflects individual subjective assessment of timelines, alignment difficulty, governance, failure modes.

The metric has evolved from informal forum slang into a serious metric used by researchers, policymakers, industry leaders.

Why It’s Inherently Uncertain

Three structural challenges:

  1. No historical data — unlike most risk assessments, no empirical base rate exists
  2. Reliance on theoretical arguments and expert judgment about scenarios that have never occurred
  3. No standard methodology — definitions of “doom” vary (extinction vs. disempowerment vs. catastrophe), timeframes vary, base assumptions vary

The Range of Expert Estimates

A 2023 survey: AI researchers’ mean estimate of extinction risk in 100 years = 14.4%. Individual estimates span almost the entire probability range:

ResearcherP(doom)
Roman Yampolskiy99.9%
Eliezer Yudkowsky>95%
Dan Hendrycks>80%
Holden Karnofsky50%
Paul Christiano46%
Dario Amodei10–25%
Yoshua Bengio20%
Geoffrey Hinton10–20%
Elon Musk10–30%
Vitalik Buterin10%
Yann LeCun<0.01%
Marc Andreessen0%

This appendix is essentially a citable concern-level snapshot — useful as policy-and-advocacy reference. The key analytical insight is the substantial probability mass that knowledgeable experts place on catastrophic outcomes — including those who built the systems creating these risks.

Caveats

  • Many experts don’t specify timeframes, making comparisons difficult
  • “Doom” is defined inconsistently (extinction vs. permanent disempowerment vs. civilizational catastrophe)
  • Estimates are highly sensitive to assumptions about timelines, alignment difficulty, and institutional response
  • Subjective ≠ arbitrary — these are inputs to prioritization and policy, not pretensions to objectivity

Connection to Wiki

This appendix grounds the wiki’s existing positions: