Research

Lab Mission

As you go through your day, you are effortlessly interacting with other people by building living simulations of who they are and how they think. This is the capacity that allows people to think together, to invent futures, coordinate at scale, and build knowledge that grows over generations. Our social intelligence doesn’t just make us the most sophisticated species on earth, it is also at the core of the human experience. The most important parts of our lives are defined by our relationships, communities, and the ways in which we become part of something larger than ourselves.

Understanding how we do this is one of the biggest open problems in science. Our lab’s mission is to build a computational theory of how minds understand each other by answering four foundational questions: How does the mind model other minds? That is, how do minds represent mental states, cognitive processes, and social roles that structure everyday life? How do we build accurate models of each other? To build a good model of another person, we need to look beyond how they look and act. We also consider the traces they leave on the world and we even use properties of our own minds to guess how theirs works. How is all of this processed and integrated? How does it develop and vary? While many social capacities appear early in childhood and may be universal, much is shaped by experience and culture. By studying variability, we hope to reveal which parts of the architecture are built-in vs. tuned by the environment. What does it make possible? What unique abilities does this give us that other creatures lack?

Our lab pursues each question through experiments to discover new phenomena, theory building to explain human thinking, and computational models that formalize these ideas and make new testable predictions. We believe that if we truly understand social intelligence, we should be able to characterize every component, build it, and replicate human intuitions.

Our vision is a future where understanding how minds understand each other becomes a pillar of science and society. Social cognition is the foundation on which nearly everything distinctly human is built: communication, cooperation, culture, and morality. None of these can be fully understood without first capturing how people understand each other. For society, we envision a future where intelligent machines think how we think and value what we value; a world where we can treat conditions where social cognition fails; and societies where differences in how we see the world don’t harden into fracture. Humanity’s greatest achievements—landing on the moon, decoding the genome, building megacities of millions—happened because minds understood each other well enough to act as one. In our most challenging times, humanity’s success, and all of the things we have yet to create, depend on our social skills and what they give us when they work: each other.

Research summaries

Click below to find short summaries on research we've done.

What is the connection between Theory of Mind and language?

Language use is a fundamentally social activity, but classical work suggested the connection between language and social cognition is slow, effortful, and error-prone. Our work with Paula Rubio-Fernandez has been challenging this view. We have found that people make real-time social inferences as speech unfolds in real time [2020, PNAS], and people can extract high-resolution mental state information from minimal linguistic cues, matching the precision of mental-state inferences from physical action [2021, Science Advances]. On the production side, we have found that even the basic linguistic choices reveal social reasoning operating at the word level, like when to say this versus that [2024, PNAS], when to use adjectives [2020, JEP:General], or how much to say to jog someone's memory [2025, CogSci]. Consistent with this, we our computational models where social cognition is embedded in language outperform standard models and explain cross-linguistic variation, including why Spanish speakers use fewer adjectives than English speakers [2022, Psych Review]. We recently published a theory paper that reviews all these findings (along with other work), to argue that everyday communication is sustained by continuous word-by-word mind-tracking, rather than secondary language-independent pragmatic reasoning [2025, TiCS].

What is The Computational Basis of Theory of Mind?

What computations and representations allow humans to reason about other minds? Our work has helped advanced the view that people infer mental states by inverting a generative model of rational action [2017, NHB]. We have proposed that this is accomplished through a good-enough simplified model of other minds that treats others as if they are utility maximizes who trade off costs and rewards when forming plans [2020, Cog. Psychology]. This approximation gets other people's knowledge and intent broadly right, making their behavior interpretable and predictable. Empirically, we have found that this system permeates social reasoning from early in childhood, helping us infer competence [2015, Cognition], preferences [2017, JEP:General], and knowledge [2021, Child Dev.], as well as guiding how we understand language [2019, Child Dev.], communicate with others [2019, NHB] and make moral judgments [2015, Psych. Science]. You can find a review paper here [2016, TiCS], and a modeling tutorial here.

More recently, we have been working on the idea that Theory of Mind is more than an ability to attribute beliefs and desires, and includes an ability to reason about cognitive processes themselves. Based on this, we are working on next-generation Theory of Mind frameworks track capture how people infer and track the cognitive processes happening in other minds, like inferring when people are reasoning, hesitating, remembering something, or getting distracted [2023, CogSci; 2021, CogSci]. We've found that this capacity is connected to first-person cognitive processes [2024, CogSci], and we've shown that young children also have expectations about cognitive processes including the dynamics of knowledge [2023, CogDev] and attention [2024, CogSci].

The Social Representation of the Physical World

How pervasive is social reasoning? Our work suggests that people use social reasoning and Theory of Mind not only when observing agents, but even when looking at objects and physical scenes. People can detect indirect of agency in their environment [2021, JOV] and use it to reconstruct happened [2022, JEP:General], and even young children can do this [2021, CogSci]. We have also shown that this capacity allows people to use the physical world itself as a channel for communication, allowing people across cultures to indirectly communicate through objects [2023, Cognition]. This work suggests that social reasoning is not only a system for extracting mental states from people's observable action, but also from how people shape their environments [2024, Current Dir. in Psych. Science; 2025, BBS Commentary].

Using Theory of Mind and Meta-Cognition for Social AI

How can people's capacity to model minds guide the design of socially competent and safe AI? Our computational models of human Theory of Mind offers a path for building machines that understand behavior through structured mental-state representations rather than pattern recognition alone [2019, Curr. Op. in Beh. Sci. Special issue on AI]. In parallel, we have also shown how meta-cognitive architectures, where systems reason about themselves, can give vision models the ability to self-discover their own hallucinations and learn when to not trust what they see [2020, NeurIPS SVRHM; 2024, UAI] (We also think this is a key capacity for human intelligence; [2024, BBS Commentary]. By contrast, we recently found that large language models have sophisticated social mimicry but lack the signatures of abstract model-based representations of other minds [2025, AAAI AI-ToM]. These projects shape how we think about broader problems in alignment, thinking about how architectures inspired by biological intelligence can constraint the design of robust and cooperative AI [2025, arXiv], and how robust social intelligence needs to scale beyond individual agents to institutions [2025, arXiv].

Making Our Minds Understood

How are people to skilled at building models of what is happening in other people's minds? We think part of the reason is because people act in ways that make themselves easy to understand--a capacity we call mindsharing. The basic idea is that people's behavior is designed to support mental-state inference. For example, people prefer taking movements that more clearly reveal their goals [2023, Cogsci]. They also use interjections (words like oops!) to help other people know what is happening [2025, Cogsci], and this is a key principle behind human gesture [2022, Nat. Comms.]. We have more findings in this area coming soon!

The Institutional Stance

One puzzle in mental-state inference is that reasoning about other minds is computationally costly, but people can easily navigate complex social environments. In collaboration with Dr. Dunham's lab, we are developing the idea that humans have an institutional stance: a system that co-evolved with Theory of Mind, to interpret behavior in terms of structures systems of roles rather than representations of mental states [in press, BBS]. Precursors of the institutional stance explain how other species coordinate without rich models of other minds. But the institutional stance is unique to humans in its generative capacity, allowing us to create arbitrary institutional structures that led to an explosion of role-based systems in modern societies (and some aspects are embedded in language itself!). In recent work we've found that role representations allow people to form instant expectations about what others will do, what they know, and who else could serve the same function [2024, CogSci]. We are also modeling how people infer social structures from sparse observations [2022, CogSci], and how we separate role-based sources of behavior from genuine mental states [2025, CogSci]. We have a lot of active work in this area, so more coming soon!

The Development of Social Cognition

How does social cognition develop and what does this tell us about the mind more broadly? Our work suggests that children can reason about others' goals and desires from very early in life [see 'Computational Basis of Theory of Mind' section above; 2015, Cognition; 2017, JEP:General; 2021, Child Dev.; 2019, Child Dev.; 2019, NHB; 2015, Psych. Science; 2016, TiCS], but knowing how to infer beliefs and knowledge takes longer. We have found that children as young as four years old already have a structured causal model that connects knowledge to action [Aboody et al., Cognition], but they initially rely on coarse heuristics to infer mental states and don't show probabilistic reasoning characteristic of adult cognition until around age six [2025a, Child. Dev; 2025b, Child. Dev]. Interestingly, adult belief inference is graded and quantitative [2025, Cognition; 2017, NHB], but not when these inferences requires representing broad hypothesis spaces [2023, Cognitive Science]. We have recently proposed a broader framework that suggests that developing social cognition involves learning to build restricted-scope models of other minds—simplified but useful representations that balance efficiency and expressive power, which we hypothesize emerge, in large part, through parent-child conversations [2025, Ann. Rev. of Dev. Psych.].

CSCL dissertations

Amanda Royka

Monkeys and Mindsharing: An Exploration of the Unique Power of Human Theory of Mind.

By Dr. Amanda Royka.

Background images by Miguel Alcântara, Liam Burnett-Blue, Yuheng Ouyang, Charlein Garcia, Rob Curran, CHUTTERSNAP, Alfauzan Nuryadin, and Susan Q Yin.