Machine Psychology
Abstract: LLMs show increasingly advanced emergent capabilities and are being incorporated across various societal domains. Understanding their behavior and reasoning abilities therefore holds significant importance. We argue that a fruitful direction for research is engaging LLMs in behavioral experiments inspired by psychology that have traditionally been aimed at understanding human cognition and behavior. In this article, we highlight and summarize theoretical perspectives, experimental paradigms, and computational analysis techniques that this approach brings to the table. It paves the way for a "machine psychology" for generative AI that goes beyond performance benchmarks and focuses instead on computational insights that move us toward a better understanding and discovery of emergent abilities and behavioral patterns in LLMs. We review existing work taking this approach, synthesize best practices, and highlight promising future directions. We also highlight the important caveats of applying methodologies designed for understanding humans to machines. We posit that leveraging tools from experimental psychology to study AI will become increasingly valuable as models evolve to be more powerful, opaque, multi-modal, and integrated into complex real-world settings.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
What is this paper about?
This paper introduces a new idea called “machine psychology.” It suggests using tools and experiments from human psychology to study how LLMs—AI systems like ChatGPT—behave, make decisions, and solve problems. Instead of looking inside the AI’s “brain” (its code and neural layers), the paper recommends watching how it responds to different questions and situations, just like psychologists do with people.
What questions does the paper try to answer?
The paper asks simple but important questions:
- Can we test AIs with the same kinds of experiments used on humans to see what they can do?
- What kinds of human-like behaviors do AIs show, like reasoning, bias, creativity, or moral judgment?
- How should we design fair, reliable tests for AIs so we don’t accidentally trick them or get misleading results?
- How should we interpret AI behavior without pretending they think or feel exactly like humans?
How did the researchers approach the problem?
The paper is a “concept” paper. It doesn’t run one big experiment; instead, it:
- Brings together many studies where AIs took classic psychology tests (like puzzles, moral questions, or creativity tasks).
- Proposes a set of testing rules and best practices to make machine psychology studies fair and dependable.
- Explains how to make sense of AI behavior using everyday language while staying careful about what we claim.
Here’s how machine psychology works in practice:
- Treat the AI like a study participant: You give it a “prompt” (a question or story), and it gives an answer.
- Use two kinds of methods:
- Self-report: Ask the AI questions (like a survey) and record its answers.
- Observational: Give it tasks or scenarios and judge its behavior based on what it does.
- Focus on inputs and outputs: Instead of peeking inside the AI, watch how different prompts lead to different responses and look for patterns.
To make these tests strong and fair, the paper recommends:
- Avoid training data contamination: Change the wording of famous tests so the AI isn’t just repeating something it saw during training.
- Multiply prompts: Try many versions of the same task because small wording changes can shift the AI’s answer.
- Control for technical biases: AIs can favor common words or the last option they saw—design prompts to reduce these effects.
- Improve reasoning: Ask the AI to “think step by step” (chain-of-thought), break big problems into smaller parts, or use multiple-choice formats to boost accuracy.
- Set the right parameters: Adjust settings like “temperature” (which controls randomness) and test with the best available model.
- Evaluate carefully: Use simple, consistent ways to check answers, and rely on human reviewers when outputs are long or complex.
What did they find?
By looking across many studies, the paper shows that modern LLMs often display human-like patterns:
- Reasoning and biases: They can solve logic problems but also make the same kinds of mistakes humans do (like being influenced by how a question is framed).
- Theory of mind: Some models can understand what others know or believe in stories—like passing false-belief tests used with children.
- Morals and values: AIs can answer moral questions, show certain value preferences, and sometimes show “moral disengagement” (excusing harmful actions).
- Creativity: On tasks that measure creativity (like finding many uses for a simple object), some AIs perform similarly to humans.
- Personality-like styles: AIs can appear more “agreeable” or “extraverted” depending on how they’re trained.
- Learning from examples: LLMs are good at “few-shot learning”—improving at a task after seeing a few examples.
- Working in groups: Multiple AIs talking to each other can show new dynamics (like debating or collaborating), which opens up “group psychology” for machines.
Most importantly, machine psychology helps discover “emergent” abilities—skills that don’t show up on traditional AI benchmarks but appear in richer, more realistic tests.
Why does this matter?
- Safety and alignment: Understanding AI behavior helps us predict and prevent harmful actions, especially as AIs become part of search engines, assistants, and decision tools.
- Better testing: Machine psychology complements standard benchmarks by checking higher-level skills like judgment, values, and social behavior.
- Clearer explanations: Watching behavior rather than code can make AI more understandable to users, policymakers, and scientists.
- Planning for the future: As AIs become multimodal (able to handle text, images, and tools) and more powerful, we’ll need these methods to track changes over time and spot new abilities early.
How should we think about AI behavior?
The paper warns us to be careful with language. It’s tempting to say an AI “thinks,” “feels,” or “wants” things, but its inner workings aren’t like a human brain. Still, using everyday psychological terms can help explain what the AI is doing in a way people understand. The key is to:
- Describe patterns honestly: Say what the AI does and when it does it.
- Avoid overclaiming: Don’t pretend AIs have human-like minds or emotions.
- Use “thick” descriptions when helpful: It’s okay to use words like “reasoning” or “creativity” to explain behavior, as long as we remember they’re analogies, not exact matches.
Final takeaway
Machine psychology is a new, practical way to study AI behavior using tools from human psychology. It helps us spot surprising abilities, understand risks, and build better, safer AI systems. As LLMs grow more capable and interact more with the world, these methods will become essential for guiding how we design, test, and trust AI.
Collections
Sign up for free to add this paper to one or more collections.