This shows a person in a dark room and golden sound waves.
Researchers discovered that expert echolocators improve their spatial accuracy by "stacking" information from repeated mouth clicks. Credit: Neuroscience News

The Brain Stacks Sound Information to Navigate the Dark

Summary: Some blind individuals possess a “superpower” known as echolocation, using the echoes of their own mouth clicks to navigate the world. A new study published has finally mapped how the brain processes these sounds.

By comparing expert echolocators to sighted individuals in total darkness, the team found that the brain doesn’t just “hear” an objectโ€”it accumulates information across a sequence of clicks. Each additional click acts like a brushstroke, building a high-resolution mental representation of the surroundings in real-time.

Key Facts

  • Expert Precision: Four blind expert echolocators significantly outperformed 21 sighted individuals at identifying object locations in a pitch-black room.
  • The “Build-Up” Effect: Neural activity in the brain strengthens with every successive click. The more clicks an expert makes, the more accurate their “vision-by-sound” becomes.
  • Information Summation: Lead researcher Haydee Garcia-Lazaro found that the brain uses “summation” to combine repeated sound information into a stable spatial map.
  • Behavioral Correlation: The accuracy of object location improved linearly with the number of self-generated mouth clicks, proving that echolocation is an active, iterative process.
  • Future Training: The study suggests that echolocation is a skill that can be trained in both blind and sighted individuals by engaging specific neural pathways used for spatial representation.

Source: SfN

Some blind people use returning echoes from theirย ownย mouth clicks toย perceiveย externalย surroundings,ย orย echolocation.ย 

New fromย eNeuro, Haydee Garcia Lazaro and Santani Teng, from Smithโ€“Kettlewell Eye Research Institute,ย explored how the human brain creates representations of external surroundings using echolocation.ย 

The researchers first discovered thatย fourย blind individualsย comfortable withย using echolocationย couldย identifyย objectย location better thanย 21 peopleย with vision intactย in a dark room. Accuracy at using echolocation improved with moreย self-generatedย mouth clicksย inย theseย expert echolocators.ย 

The researchers also linked neural activity in the brain to the ability of blind individuals toย determineย object location. This activity, alongsideย behavioral measures,ย strengthened across click sequences, leading to moreย accurateย object location.ย 

Says Garcia-Lazaro, โ€œBasically, we found that, in some experts, there appears to be a summation, or accumulation, of information in the brain that builds up across clicks about object location.โ€ย 

According to the researchers, this workย shows how the brain usesย repeated sound information to createย representations of the environment in the absence of vision.ย 

Garcia-Lazaro expresses excitement aboutย the nextย steps stemming from this work, includingย determiningย what makes blind people adept at echolocationย and training people withย and withoutย sight to engage their echolocation ability.ย 

Key Questions Answered:

Q: Do echolocators actually “see” the objects in their minds?

A: In a way, yes. While they are using sound, the brain often repurposes the visual cortex to process this spatial information. This study shows that the brain is essentially building a “3D model” click-by-click, which allows experts to sense the distance, size, and even the texture of objects.

Q: Why don’t sighted people have this ability?

A: We actually do have a “latent” version of it! Sighted people in the study were able to do it, just much less accurately. Expert echolocators have “neuroplastic” brains that have been tuned through years of practice to prioritize these subtle echoes, whereas sighted brains usually ignore them in favor of light.

Q: Can anyone learn to echolocate?

A: The researchers believe so. Because they identified a specific neural “summation” process, it opens the door for training programs. Just as you can learn to play an instrument, you can likely train your brain to “listen” for the build-up of spatial data from your own clicks.

Editorial Notes:

  • This article was edited by a Neuroscience News editor.
  • Journal paper reviewed in full.
  • Additional context added by our staff.

About this visual and auditory neuroscience research news

Author:ย SfN Media
Source:ย SfN
Contact:ย SfN Media โ€“ SfN
Image:ย The image is credited to Neuroscience News

Original Research:ย Closed access.
โ€œNeural and Behavioral Correlates of Evidence Accumulation in Human Click-Based Echolocationโ€ by Haydรฉe G Garcรญa-Lรกzaroย andย Santani Teng.ย eNeuro
DOI:10.1523/ENEURO.0342-25.2026


Abstract

Neural and Behavioral Correlates of Evidence Accumulation in Human Click-Based Echolocation

Echolocation enables blind individuals to perceive and navigate their environment by emitting clicks and interpreting their returning echoes. While expert blind echolocators demonstrate remarkable spatial accuracy, the behavioral and neural mechanisms by which spatial echoacoustic cues are combined across repeated samples remain less explored.

Here, we investigated the temporal dynamics of spatial information processing in human click-based echolocation using EEG. Blind expert echolocators (n=4, all males) and novice sighted participants (n=21, 12 males) localized virtual spatialized echoes derived from realistic synthesized mouth clicks, presented in trains of 2โ€“11 clicks. Behavioral results showed that blind expert echolocators significantly outperformed sighted controls in spatial localization.

For these experts, localization thresholds decreased as the number of clicks increased, a pattern consistent with cumulative integration of spatial information across repeated samples. EEG decoding analyses revealed reliable neural discrimination of echo laterality from the first click that correlated with overall spatial localization performance.

Across successive clicks, neural responses evolved systematically, reflecting sequence-positionโ€“dependent changes in neural dynamics. EEG trial-level modeling further allowed us to distinguish accumulation-consistent decision readout policies from alternative repetition-based accounts, revealing individual differences in decision policies among expert echolocators.

These findings provide, to our knowledge, the first fine-grained account of the temporal neural dynamics supporting human click-based echolocation, directly linked to behavioral performance across multiple samples.

They reveal how, in expert echolocators who successfully performed the task, successive echoes are progressively integrated into coherent spatial representations, demonstrating adaptive sensory processing in the absence of vision.

Join our Newsletter
I agree to have my personal information transferred to AWeber for Neuroscience Newsletter ( more information )
Sign up to receive our recent neuroscience headlines and summaries sent to your email once a day, totally free.
We hate spam and only use your email to contact you about newsletters. You can cancel your subscription any time.