AGI metrics — SR2025 Agenda Snapshot
One-sentence summary: Evals with the explicit aim of measuring progress towards full human-level generality.
Theory of Change
Help predict timelines for risk awareness and strategy.
Broad Approach
behavioural
Target Case
mixed
Key People
CAIS, CFI Kinds of Intelligence, Apart Research, OpenAI, METR, Lexin Zhou, Adam Scholl, Lorenzo Pacchiardi
Funding
Leverhulme Trust, Open Philanthropy, Long-Term Future Fund
Estimated FTEs: 10-50
Critiques
Is the Definition of AGI a Percentage?, The “Length” of “Horizons”
See Also
Outputs in 2025
5 item(s) in the review. See the wiki/summaries/ entries with frontmatter agenda: agi-metrics (these were generated alongside this file from the same export).
Source
- Row in
shallow-review-2025/agendas.csv(name = AGI metrics) — Shallow Review of Technical AI Safety 2025.
Related Pages
- ai-safety
- ai-safety
- capability-evals
- ai-deception-evals
- ai-scheming-evals
- autonomy-evals
- other-evals
- sandbagging-evals
- self-replication-evals
- situational-awareness-and-self-awareness-evals
- steganography-evals
- various-redteams
- wmd-evals-weapons-of-mass-destruction
Sources cited
Primary URLs harvested from this page’s summary references. Auto-generated by scripts/backfill_citations.py; edit by re-running, not by hand.
- Summary: AI Safety (Wikipedia) — referenced as
[[ai-safety]]