National AI Governance

National AI governance refers to country-level regulatory frameworks for AI. Major powers have adopted distinct regulatory philosophies that reflect their broader political values — creating a fragmented global landscape that complicates international coordination. The AI Safety Atlas (Ch.4.8 appendix) provides the canonical comparative analysis.

Three Regulatory Philosophies

Across 30+ countries with AI strategies, three dominant patterns:

Development-led

Examples: China, South Korea Approach: state directs resources toward infrastructure and national missions Logic: AI as strategic asset; government coordinates investment, talent, infrastructure

Control-oriented

Examples: EU, Norway, Mexico Approach: legal standards, ethics oversight, risk monitoring Logic: AI as risk source; government protects citizens through binding rules

Promotion-focused

Examples: US, UK, Singapore Approach: state acts as enabler with minimal regulatory constraints Logic: AI as innovation engine; government clears barriers, lets market sort outcomes

Country Case Studies

European Union — Rights-Centric Horizontal

  • EU AI Act (March 2024) — world’s first comprehensive AI legal framework
  • Risk-tier classification — unacceptable / high / limited / minimal
  • Implementation timeline — Feb 2025 prohibitions; Aug 2025 GPAI obligations
  • Enforcement — European AI Office; fines up to 3% global turnover
  • Brussels Effect — see brussels-effect
  • See eu-ai-act for detail

United States — Geopolitical Competition

  • Major shift post-2024 — Trump’s EO 14179 revoked Biden’s October 2023 AI safety EO
  • Current direction — federal agencies remove barriers; AI must be free of “ideological bias or engineered social agendas”
  • Distinctive lever — semiconductor export controls (NVIDIA, AMD, Intel under US jurisdiction)
  • Constraint — congressional gridlock → executive orders only
  • State-level activity — California SB 1047 vetoed (Sept 2024); SB 53 whistleblower protection introduced

China — Vertical Iterative

  • Distinctive approach — targeted regulations for specific AI domains, not horizontal framework
  • Key regulations — Algorithmic Recommendation Provisions (2021, world’s first algorithm registry); Deep Synthesis Provisions (2022); Generative AI Interim Measures (2023); AI-Generated Content Labeling (2025)
  • Enforcement — Cyberspace Administration of China with broad discretion
  • Required values — “socialist core values”; algorithm registry
  • Inward focus — primarily regulates Chinese organizations; Western labs (OpenAI, Anthropic) don’t actively serve Chinese consumers due to censorship-compliance refusal
  • Building toward — proposed comprehensive Artificial Intelligence Law

Comparative Summary

JurisdictionPrimary FocusMechanism
EUIndividual rights protectionHorizontal comprehensive framework
USGeopolitical competition with ChinaHardware export controls + executive orders
ChinaSocial control + government value alignmentVertical iterative algorithm-focused regulation

Why Fragmentation Is the Problem

Each approach reflects its country’s broader values. None is straightforwardly “wrong.” But the combination is fragmented governance that:

  • Complicates international coordination
  • Creates regulatory arbitrage incentives
  • Makes coordinated red lines harder to negotiate
  • Pushes development toward least-regulated jurisdictions (race-to-bottom)

The Atlas’s implicit argument: national governance is necessary but insufficient — international coordination must bridge the divergent philosophies.

Lessons from Nuclear Safety

The Atlas points to nuclear safety regulation as a blueprint:

  • Standardized safety frameworks
  • Independent supervision mechanisms
  • Regular protocols and incident-response exercises
  • Information sharing across industries

These patterns inform the proposed AI governance architecture but require translation to AI’s distinctive properties (proliferation, emergent capabilities, etc.).

Connection to Wiki

Sources cited

Primary URLs harvested from this page’s summary references. Auto-generated by scripts/backfill_citations.py; edit by re-running, not by hand.