Join the EulerFold community
Track progress and collaborate on roadmaps with students worldwide.
Generative Agents
Park, J. S., O'Brien, J. C., Cai, C. J., Morris, M. R., Liang, P., & Bernstein, M. S. (2023). Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442.
Read Original Paper
In 2023, the 'Generative Agents' paper from Stanford and Google introduced a way to create believable digital characters that can plan their days, form relationships, and coordinate activities autonomously. While previous non-player characters (NPCs) in games relied on rigid scripts or simple state machines, these agents used large language models to simulate the complexity of human life. The researchers populated a sandbox world with 25 agents and observed how individual actions coalesced into social dynamics. It was a shift from programming behaviors to architecting memories.
The Ranked Memory Stream

The generative agent architecture: perceiving, remembering, reflecting, and planning in a continuous loop.
Generative Agents established a framework for believable digital characters by architecting a persistent memory stream that informs every agentic decision. To manage the finite context of a language model, the researchers developed a ranking function that retrieves memories based on a weighted sum of their recency, importance, and semantic relevance to the current situation. This long-term record is periodically synthesized through a "reflection" mechanism, where the model extracts high-level insights from raw observations and stores them back as abstract concepts that define the agent's identity. This shift from programmed state machines toward a recursive architecture of memory and reflection proved that complex social behaviorsâsuch as information diffusion and long-term planningâemerge naturally when an agent is given the ability to recall and reason about its own history.
The Reflection mechanism
To enable agents to generalize from raw observations, the researchers introduced a 'Reflection' mechanism. Without reflection, an agent might remember 'Sam is eating breakfast' multiple times but never realize 'Sam likes to sleep late.' This process is triggered when the importance scores of recent memories exceed a threshold. The agent generates salient questions based on its history, retrieves related memories, and then prompts the model to extract high-level 'insights' or reflections. These reflections are stored back in the Memory Stream as new experiences, effectively creating a deeper layer of personal identity. It proved that intelligence requires a periodic pause to synthesize raw data into abstract concepts, mimicking the human process of self-discovery.
Recursive Planning and Action
Long-term coherence in agent behavior was achieved through a top-down planning system. Agents first generate a broad schedule for the day, which is then recursively refined into detailed 15-minute blocks. This allows an agent to maintain a consistent goalâlike going to workâwhile remaining reactive to immediate environment changes. For instance, if an agent perceives a fire or starts a conversation, it can choose to 're-plan,' adjusting its entire future schedule based on the new context. This approach proved that 'believability' in a simulation is a product of balancing rigid long-term plans with flexible, moment-to-moment reactions. It suggested that the future of agentic systems lies in the ability to bridge the gap between high-level intent and low-level execution.
Dive Deeper
Generative Agents Demo
Stanford ⢠docs
Explore ResourceGenerative Agents Paper on arXiv
arXiv ⢠article
Explore Resource