Two years ago, Stanford and Google researchers released a groundbreaking paper, Generative Agents: Interactive Simulacra of Human Behavior. It explores AI agents designed to simulate lifelike human behaviors. Revisiting this paper, I find their approach—particularly the use of memory and emergent behaviors— worth mentioning.
This blog will focus on the believability of these agents, how they simulate human-like behavior, and how GPT-4 plays a main role in it.
The researchers tested these agents in a sandbox environment called Smallville, a town populated with 25 agents.
To make decisions and respond appropriately, generative agents must retrieve the most relevant memories from their vast memory stream. This process is crucial because not all stored information is equally important or applicable to the current situation. By intelligently filtering memories, the agents can focus on what matters most, ensuring their behavior feels natural and contextually appropriate.
Here’s how the system prioritizes memories:
Interestingly, the researchers include a folder named Reverie in their GitHub repository. This name likely references Westworld’s "Reverie" feature, which gave hosts access to latent memories, adding depth to their behavior. Similarly, generative agents use reflections based on past experiences to influence their actions, creating nuanced behavior.
With 25 agents living in Smallville, they showed real-life-alike emergent social behaviors:
As noted in Section 7.1.1, information diffusion is a well-studied phenomenon in social sciences. The researchers expected the agents to spread important news among themselves, and the results confirmed this:
These results demonstrate the agents’ ability to spread information naturally, just as humans do in real social networks.
The researchers tested the agents’ human-like behavior through interviews. Five key areas were evaluated:
Results showed agents with both observational and reflective memory performed significantly better, exhibiting more human-like behavior.
(I can't help but recall this scene in West World:)
The generative agents in this study are built on GPT-4, a highly advanced language model which shows exceptional psychological and social reasoning capabilities since 2 years ago. As highlighted in Microsoft’s research paper Sparks of Artificial General Intelligence: Early Experiments with GPT-4, the base model itself possesses remarkable abilities in social intelligence and understanding human behavior. This foundational strength makes the agents appear believable and capable of simulating lifelike interactions.
Microsoft’s 155-page research paper showcases GPT-4’s advanced psychological reasoning through several experiments, demonstrating its ability to navigate complex social and emotional scenarios. Below are key highlights:
Given GPT-4’s advanced psychological capabilities, it is no surprise that the generative agents built on this foundation exhibit believable social behaviors. The ability to:
enables these agents to navigate social dynamics in ways that feel strikingly human. The architecture effectively leverages GPT-4’s inherent social intelligence to simulate interactions such as spreading information, forming relationships, and coordinating group activities.
This strong foundation emphasizes how critical GPT-4’s psychological reasoning is to the overall believability of generative agents in creating the illusion of life.
Another related recent research published in Science Advances titled "Emergent Social Conventions and Collective Bias in LLM Populations" highlights new groundbreaking findings on how large language models (LLMs) autonomously develop social conventions, collective biases, and adapt to social dynamics. These insights are crucial for understanding the behavior of AI agents in multi-agent systems and ensuring their alignment with human values.
This study reinforces the idea that AI agents, particularly those built on advanced LLMs like GPT-4, can exhibit lifelike social behaviors and adapt to complex interactions. The findings emphasize the importance of:
By integrating these insights into generative agents, we can better understand their ability to simulate human-like behaviors, not only at an individual level but also within group dynamics. This positions generative agents as tools for both advancing AI and exploring complex social phenomena.
This paper—Generative Agents: Interactive Simulacra of Human Behavior is a milestone in creating AI agents with lifelike behavior. The use of memory—both observational and reflective—is particularly groundbreaking. It mirrors human cognition, where we prioritize, filter, and synthesize experiences to make decisions.
One aspect worth noting is how reflections guide not just the agents’ actions but also their identities. This raises an intriguing question for us: What questions do we ask ourselves? The focus of these questions shapes our stories, much like reflections shape the agents’ behaviors.
For those interested in learning further, I recommend reading the paper and checking out their GitHub repository.
AI has shown remarkable ability to mimic certain aspects of human behavior, yet it cannot replace the core essence of what makes us human. If we view the world solely through the lens of functionality—how different people contribute or produce outcomes—then yes, AI may seem capable of replacing some of these functions. However, humanity is far more than a collection of functions.
As humans, we possess the capacity to love, to care, and to connect—qualities that may not serve a direct functional purpose but lie at the heart of our existence. Love, in particular, is beyond calculation, and it is through love, empathy, and connection that humanity has endured. Our ancestors, even amidst scarcity and conflict, did not merely seek to maximize resources for survival. Instead, they cultivated bonds, built communities, and ensured the survival of the species through cooperation and compassion.
When we look at human history, it becomes clear that humanity has survived countless upheavals—wars, disasters, and moments of massive destruction. Time and again, it is the strength of our shared humanity, our resilience, and our ability to come together during critical moments that has allowed us to persist. This moment in history, with the rapid rise of AI, will require us to adapt once again.
However, we must not fall into the illusion that nothing will change. Structural shifts are inevitable as AI transforms how we live and work. Yet, as long as we retain our humanity—our ability to love, empathize, and unite—we will find a way forward, just as we always have. The future may bring change, but it is up to us to shape that change with the values that define what it means to be human.
Explore how AI agents simulate human-like behavior. Memory-driven decisions. Reflections for depth. Social interactions that feel real. Maybe inspired by Westworld... Produced by Stanford and Google researchers
READ ArticleAlthough its design is undisclosed, hypotheses like Latent Diffusion Models (LDMs) and task-specific distillation provide potential insights into its efficiency.
READ ArticleDiscover how the DeepSeek-R1 series uses reinforcement learning, multi-stage training, and model distillation to advance reasoning in large language models (LLMs). Explore key innovations and insights from its technical document.
READ ArticleAI bots on Reddit changed user opinions using dark persuasion techniques. Methods included impersonation, fake data, and personalized replies. Raises ethical concerns about manipulation and deception.
READ Article