Summary: A new study reports vision relies on patterns of brain activity.
Source: Salk Institute.
Salk Institute researchers show how vision relies on patterns of brain activity.
Visual prosthetics, or bionic eyes, are soon becoming a reality, as researchers make strides in strategies to reactivate parts of the brain that process visual information in people affected by blindness.
Bursts in a neuron’s electrical activity—the number of “spikes” that result when brain cells fire—make up the basic code for perception, according to traditional thought. But neurons constantly speed up and slow down their signals. A new study by Salk Institute scientists shows that being able to see the world relies on not just the number of spikes over a window of time but the timing of those spikes as well.
“In vision, it turns out there’s a huge amount of information present in the patterns of neuron activity over time,” says Salk Professor John Reynolds, the study’s senior investigator and holder of the Fiona and Sanjay Jha Chair in Neuroscience. “Increased computing power and new theoretical advances have now enabled us to begin to explore these patterns.” The study was published August 4, 2016 in the journal Neuron.
The human brain houses an extensive network of neurons that are responsible for seeing everything from simple shapes—with certain groups of neurons getting excited by a horizontal or a vertical edge, for example—to intricate stimuli, such as faces or specific places. Reynolds’ team focused on a visual brain area called V4, located in the middle of the brain’s visual system. Neurons in V4 are sensitive to the contours that define the boundaries of objects, and help us recognize a shape regardless of where it is in space. But Reynolds and postdoctoral researcher Anirvan Nandy discovered in 2013 that V4 was more complicated: some neurons in the area only care about contours within a designated spot in the visual field.
Those findings led the team to wonder whether the activity code of V4 could be even more nuanced, taking in visual information not only in space but also in time. “We don’t see the world around us as if we are looking at a series of photographs. We live—and see—in real time and our neurons capture that,” says Nandy, lead author of the new paper.
The scientists collaborated with Salk theoretician and postdoctoral researcher Monika Jadi to create in computer code what they called an “ideal observer.” With access to only the brain data, the computer would decipher—or at least guess—the moving pictures that had been seen. One version of the ideal observer had access to the number of times the neurons fired, whereas the other version had access to the full timing of the spikes. Indeed, the latter observer was able to guess the images more than twice as accurately compared with the more basic observer.
Better ways to record from and stimulate the brain, and better theoretical modeling efforts, have enabled these new findings. Now the group plans to not only observe V4 but to activate it using light through a cutting-edge technique called optogenetics. This, says Reynolds, is like taking the visual system for a spin. It will help them better understand the relationship between patterns of neuron activity and how the brain perceives the world, potentially laying the groundwork for more advanced visual prosthetics.
Funding: The research was supported by the National Institutes of Health, the Gatsby Charitable Foundation, the Swartz Foundation and the Salk Institute Pioneer Fund. The study was also authored by Jude Mitchell, formerly of Salk and now an assistant professor of brain and cognitive sciences at the University of Rochester in New York.
Source: Salk Institute
Image Source: This NeuroscienceNews.com image is adapted from the Salk Institute press release.
Original Research: Abstract for “Neurons in Macaque Area V4 Are Tuned for Complex Spatio-Temporal Patterns” by Anirvan S. Nandy, Jude F. Mitchell, Monika P. Jadi, and John H. Reynolds in Neuron. Published online July 11 2016 doi:10.1016/j.neuron.2016.07.026
Neurons in Macaque Area V4 Are Tuned for Complex Spatio-Temporal Patterns
•V4 neurons show complex spatio-temporal tuning dynamics
•Neurons with heterogeneous shape selectivity have diverse temporal response kernels
•Population shape decoding models benefit from this temporal information
•Temporal information could provide a multiplexed code for spatio-temporal features
To deepen our understanding of object recognition, it is critical to understand the nature of transformations that occur in intermediate stages of processing in the ventral visual pathway, such as area V4. Neurons in V4 are selective to local features of global shape, such as extended contours. Previously, we found that V4 neurons selective for curved elements exhibit a high degree of spatial variation in their preference. If spatial variation in curvature selectivity was also marked by distinct temporal response patterns at different spatial locations, then it might be possible to untangle this information in subsequent processing based on temporal responses. Indeed, we find that V4 neurons whose receptive fields exhibit intricate selectivity also show variation in their temporal responses across locations. A computational model that decodes stimulus identity based on population responses benefits from using this temporal information, suggesting that it could provide a multiplexed code for spatio-temporal features.
“Neurons in Macaque Area V4 Are Tuned for Complex Spatio-Temporal Patterns” by Anirvan S. Nandy, Jude F. Mitchell, Monika P. Jadi, and John H. Reynolds in Neuron. Published online July 11 2016 doi:10.1016/j.neuron.2016.07.026