ARTICLE AD BOX
Researchers from Stanford, Washington University, and Google DeepMind have created AI agents that can closely mimic human behavior in social experiments.
According to the study, such simulations could serve as a laboratory for testing theories in fields such as economics, sociology, organization, and political science. The team built these agents using interview data from more than 1,000 people selected to represent the US population across age, gender, education, and political views.
![research design with human participants and simulated agents undergoing identical tests and being compared.](https://the-decoder.com/wp-content/uploads/2024/11/research-methodology-generative-agents-simulation.png)
The system works by combining detailed interview transcripts with GPT-4o. When someone queries an agent, it loads the interview transcript into the model and instructs it to imitate the person based on their responses. To create these transcripts, the researchers conducted two-hour interviews with each participant and used OpenAI's Whisper model to convert the conversations to text.
![Generative agent with memory component processes questions and interview transcripts for predictions with Expert Reflection.](https://the-decoder.com/wp-content/uploads/2024/11/generative-agent-memory-interview-reflection-diagram.png)
Interview-based agents outperform demographic agents
The research team put these AI agents through several tests to measure their ability to predict human behavior. They used questions from the General Social Survey, Big Five personality assessments, and multiple behavioral economics games.
Ad
THE DECODER Newsletter
The most important AI news straight to your inbox.
✓ Weekly
✓ Free
✓ Cancel at any time
The AI agents based on interview data predicted human GSS responses with 85% accuracy, performing significantly better than AI agents that only used basic demographic information.
![Three scatterplots compare the accuracy and correlation of different data collection methods in social surveys, personality tests, and economic games.](https://the-decoder.com/wp-content/uploads/2024/11/comparison-survey-personality-economic-accuracy-correlation.png)
The researchers ran five social science experiments with both human participants and AI agents. In four out of these five studies, the AI agents produced results that closely matched human responses. The statistical measurements showed a strong correlation between AI and human responses, with a correlation coefficient of 0.98.
![Three diagrams compare demographic parity for different survey methods based on gender, ethnicity, and political ideology.](https://the-decoder.com/wp-content/uploads/2024/11/demographic-parity-survey-comparison-analysis-1200x911.png)
The interview-based approach showed significant improvements in handling bias compared to methods using only demographics. The AI agents made more accurate predictions across different political ideologies and ethnic groups. They also showed more balanced performance when analyzing responses between various demographic categories.
Access to research data
The research team has made their dataset of 1,000 AI agents available to other scientists through GitHub. They created a two-tier access system to protect participant privacy while supporting further research. Scientists can freely access combined response data for specific tasks, while access to individual response data for open-ended research requires special permission.
This system aims to help researchers study human behavior while maintaining strong privacy protections for the original interview participants. The dataset could serve as a testing ground for theories in economics, sociology, and political science.
Recommendation