How AI “Brain States” Decode Reality

Summary: Do AI chatbots truly understand the world, or are they just repeating text? A new study suggests that AI models develop a mathematical “understanding” of real-world constraints.

By using mechanistic interpretability, essentially neuroscience for AI, researchers found that models generate distinct internal “brain states” to categorize events as commonplace, unlikely, impossible, or nonsensical. These internal maps not only mirror physical reality but also accurately reflect human uncertainty about ambiguous scenarios.

Key Facts

  • The Threshold of Understanding: An internal “world model” begins to emerge in AI systems once they reach approximately 2 billion parameters, a relatively small size compared to modern trillion-parameter models.
  • Vector Differentiation: Large models develop distinct mathematical patterns (vectors) that can distinguish between “improbable” and “impossible” events with 85% accuracy.
  • Mirroring Human Intuition: The AI’s internal states capture human-like nuance. If humans are 50/50 on whether an event (like “cleaning a floor with a hat”) is unlikely or impossible, the model’s internal probability typically reflects that same split.
  • Causal Encoding: The research suggests that by “devouring” massive amounts of text, AI models effectively reverse-engineer the causal constraints of the physical world, moving beyond simple word prediction.

Source: Brown University

Most of what AI chatbots know about the world comes from devouring massive amounts of text from the internet — with all its facts, falsehoods, knowledge and nonsense. Given that input, is it possible that AI language models have an “understanding” of the real world?

As it turns out, they do — or at least something like an understanding. That’s according to a new study by researchers from Brown University to be presented on Saturday, April 25 at the International Conference on Learning Representations in Rio de Janeiro, Brazil.

This shows a digital brain.
This work reveals evidence that language models have encoded the causal constraints of the real world in a way that predicts human judgment. Credit: Neuroscience News

The study looked under the hood of several AI language models to look for signs that they know the difference between events and scenarios that are commonplace, unlikely, impossible or downright nonsense.

“This work reveals some evidence that language models have encoded something like the causal constraints of the real world,” said Michael Lepori, a Ph.D. candidate at Brown who led the work. “Beyond just encoding these constraints, they do so in a way that is predictive of human judgments of these categories.”

Lepori’s research explores the intersection of computer science and human cognition. He is advised by Ellie Pavlick, a professor of computer science, and Thomas Serre, a professor of cognitive and psychological sciences, both of whom are faculty affiliates of Brown’s Carney Institute for Brain Science and co-authors of the research.

For the study, the researchers designed an experiment to test how language models interpret sentences describing events of varying plausibility. Some statements described commonplace scenarios: For example, “Someone cooled a drink with ice.” Some scenarios were improbable or unlikely: “Someone cooled a drink with snow.” Some were impossible: “Someone cooled a drink with fire.” Some were nonsensical: “Someone cooled a drink with yesterday.”

For each input, the researchers examined the resulting mathematical states generated inside the AI model, an approach known as mechanistic interpretability.

“Mechanistic interpretability can be appropriately characterized as something like neuroscience for AI systems,” Lepori said. “It seeks to reverse-engineer what the model is doing when exposed to a particular input. You could kind of think about it as understanding what is encoded in the ‘brain state’ of the machine.”

By comparing the differences in “brain states” generated by pairs of sentences from different categories — commonplace versus improbable, improbable versus impossible and so on — the researchers could get a sense of whether, and how well, the models internally differentiate between categories.

The experiments were repeated across several different open-source language models, including Open AI’s GPT 2, Meta’s Llama 3.2 and Google’s Gemma 2, to get a “model-agnostic” sense of how well these types of models distinguish between categories.

The study found that models of sufficient size do indeed develop distinct mathematical patterns, or vectors, that are strongly correlated with each plausibility category. The vectors could distinguish between even the most similar of categories — like improbable versus impossible events — with roughly 85% accuracy.

What’s more, Lepori says, the vectors revealed by the study are reflective of human uncertainty about which category a statement might fall into. Take the statement, “Someone cleaned the floor with a hat,” for example. When people hear that statement, they may disagree about whether it represents something that’s impossible or just unlikely. For the study, the researchers analyzed the vectors to see how ambiguous the AI systems thought these statements were, and compared that with survey results from human participants.

“What we show is that the models actually capture that human uncertainty pretty well,” Lepori said. “In cases where, say, 50% of people said a statement was impossible and 50% said it was improbable, the models were assigning roughly 50% probability as well.”

Taken together, the results suggest that modern AI language models can indeed develop an understanding of the real world that is reflective of human understanding. These vectors start to emerge in models with more than 2 billion parameters, the research found, which is fairly small compared to today’s trillion-plus-parameter models.

More broadly, the researchers say these kinds of mechanistic interpretability studies can help in developing a better understanding of what AI models know and how they came to know it.

And that, the researchers say, will help in developing smarter, more trustworthy models. 

Key Questions Answered:

Q: How can a computer know what is “impossible” if it has never been outside?

A: Through massive exposure to human language, AI identifies patterns of cause and effect. It learns that “cooling a drink with ice” is mentioned in logical, frequent contexts, while “cooling a drink with fire” appears only in contexts describing errors or fiction. This study proves the AI stores these differences as distinct mathematical categories.

Q: What is “mechanistic interpretability”?

A: Think of it as a digital MRI. Instead of just looking at the AI’s final answer, researchers look at the millions of mathematical “neurons” firing inside the model. By observing these internal states, they can see exactly how the AI is categorizing a sentence before it ever types a response.

Q: Does this mean AI is becoming sentient?

A: Not necessarily. It means the AI is building a highly accurate “internal map” of our world to predict language better. It has “understanding” in the sense that it knows the rules of our reality, but that doesn’t imply it has feelings or consciousness.

Editorial Notes:

  • This article was edited by a Neuroscience News editor.
  • Journal paper reviewed in full.
  • Additional context added by our staff.

About this AI and auditory neuroscience research news

Author: Kevin Stacey
Source: Brown University
Contact: Kevin Stacey – Brown University
Image: The image is credited to Neuroscience News

Original Research: The findings will be presented at the International Conference on Learning Representations

Join our Newsletter
I agree to have my personal information transferred to AWeber for Neuroscience Newsletter ( more information )
Sign up to receive our recent neuroscience headlines and summaries sent to your email once a day, totally free.
We hate spam and only use your email to contact you about newsletters. You can cancel your subscription any time.