LLM AGI will have memory, and memory changes alignment

Seth Herd — 2025-04-04 — LessWrong

Summary

Argues that LLM-based AGI will likely have memory systems enabling learning during deployment, and that this learning can functionally change alignment through accumulation of new beliefs and goals. Proposes empirical investigation of belief and value evolution in LLMs with memory.

Source