Summary: Researchers report on how certain activity in face cells can alter perception.
How the electrical activity of the brain gives rise to the rich world of perception.
The human brain is constantly abuzz with electrical activity as brain cells, called neurons, respond to sensory input and give rise to the world we perceive. Six particular regions of the brain, called face patches, contain neurons that respond more to faces than to any other type of object. New research from Caltech shows how perturbations in these face cells alter perception, answering a longstanding question in cognitive science.
The research was led by Doris Tsao (BS ’96), professor of biology, Tianqiao and Chrissy Chen Center for Systems Neuroscience Leadership Chair, and Howard Hughes Medical Institute Investigator. A paper describing the work appeared online in the March 13 issue of Nature Neuroscience.
“Face cells will produce the maximum response when a subject is observing faces, but they will also produce a small amount of activity when a subject is looking at round objects like an apple or a clock,” says Tsao. “There has been a long debate in cognitive science: Is the brain actually using these small responses to generate perception? Do face cells help us perceive clocks and apples?”
Tsao and her team aimed to test this by altering the perception of a trained monkey. First, the monkey was trained to look left if the two images were identical, and to look right if they were different. Then, the researchers showed the monkey two identical images of faces, but while it was viewing the second one, the researchers stimulated specific face patches. The monkey’s response changed dramatically: the animal almost always indicated that the two identical faces were different—implying that the activity of face patch neurons plays a large role in generating our perception of faces. Surprisingly, the group found that stimulating face cells also had a significant effect on the perception of certain other objects.
“There was a study a few years ago in humans, where a neurosurgeon stimulated the face area of a person, and the person exclaimed, ‘Whoa! Your face just metamorphosed,'” says Tsao. “Our study systematically perturbed each patch in the face patch network with a much larger set of stimuli, and we found that we could also perturb the perception of cartoon faces, two-toned abstract images called Mooney faces, and even apples. So we’ve shown that the class of objects that can be affected by face patch stimulation is quite large, much larger than only realistic photographs of faces. Even more surprisingly, we found that some of the objects whose percept we could alter with stimulation did not even activate the face patch we were stimulating—the face patch neurons were completely silent to these objects.”
What could explain this? The group discovered that the face patch located most posterior in the temporal lobe, thought to harbor the earliest stage of face processing, was more permissive and activated by the objects whose percept could be altered by face patch stimulation. Patches that were in a more anterior part of the temporal lobe were more specialized for faces. These findings show that while face patches play a large causal role in generating the perception of faces, these regions are also part of a complex network involved in processing a much larger class of objects. The results set the stage for future experiments to study how activity across the entire network is integrated to produce perception of an object.
The work may have applications for engineers who design algorithms to teach machines to recognize objects.
“Machine vision engineers often ask: how many specialized networks do we need for robust object vision?” Tsao says. “Do we need a special network to recognize purses, shoes, cars, food? Or can we have one general purpose network? Our results suggest that the answer is a hybrid, and this may have important engineering implications.”
Coauthors include former postdoctoral scholars Sebastian Moeller and Trinity Crapse, and current Caltech postdoctoral scholar Le Chang.
Funding: Funding was provided by the Howard Hughes Medical Institute, the National Institutes of Health, and the Humboldt Foundation.
Source: Lori Dajose – CalTech
Image Source: NeuroscienceNews.com image is credited to Courtesy of D. Tsao.
Original Research: Abstract for “The effect of face patch microstimulation on perception of faces and objects” by Sebastian Moeller, Trinity Crapse, Le Chang & Doris Y Tsao in Nature Neuroscience. Published online March 13 2017 doi:10.1038/nn.4527
The effect of face patch microstimulation on perception of faces and objects
What is the range of stimuli encoded by face-selective regions of the brain? We asked how electrical microstimulation of face patches in macaque inferotemporal cortex affects perception of faces and objects. We found that microstimulation strongly distorted face percepts and that this effect depended on precise targeting to the center of face patches. While microstimulation had no effect on the percept of many non-face objects, it did affect the percept of some, including non-face objects whose shape is consistent with a face (for example, apples) as well as somewhat facelike abstract images (for example, cartoon houses). Microstimulation even perturbed the percept of certain objects that did not activate the stimulated face patch at all. Overall, these results indicate that representation of facial identity is localized to face patches, but activity in these patches can also affect perception of face-compatible non-face objects, including objects normally represented in other parts of inferotemporal cortex.
“The effect of face patch microstimulation on perception of faces and objects” by Sebastian Moeller, Trinity Crapse, Le Chang & Doris Y Tsao in Nature Neuroscience. Published online March 13 2017 doi:10.1038/nn.4527