Using only data from an fMRI scan, researchers led by a Yale University undergraduate have accurately reconstructed images of human faces as viewed by other people.
*See notes below the article for disclaimer about this research.*
“It is a form of mind reading,” said Marvin Chun, professor of psychology, cognitive science and neurobiology and an author of the paper in the journal Neuroimage.
The increased level of sophistication of fMRI scans has already enabled scientists to use data from brain scans taken as individuals view scenes and predict whether a subject was, for instance, viewing a beach or city scene, an animal or a building.
“But they can only tell you they are viewing an animal or a building, not what animal or building,” Chun said. “This is a different level of sophistication.”
One of Chun’s students, Alan S. Cowen, then a Yale junior now pursing an advanced degree at the University of California at Berkeley, wanted to know whether it would be possible to reconstruct a human face from patterns of brain activity. The task was daunting, because faces are more similar to each other than buildings. Also large areas of the brain are recruited in the processing of human faces, a testament to its importance in survival.
“We perceive faces in a much greater level of detail than we perceive other things,” Cowen said.
Working with funding from the Yale Provost’s office, Cowen and post doctoral researcher Brice Kuhl, now an assistant professor at New York University, showed six subjects 300 different “training” faces while undergoing fMRI scans. They used the data to create a sort of statistical library of how those brains responded to individual faces. They then showed the six subjects new sets of faces while they were undergoing scans. Taking that fMRI data alone, researchers used their statistical library to reconstruct the faces their subjects were viewing.
Cowen said the accuracy of these facial reconstructions will increase with time and he envisions they can be used as a research tool, for instance in studying how autistic children respond to faces.
Chun said the study shows the value of funding research ambitions of Yale undergraduates.
“I would never have received external funding for this, it was too novel,” Chun said.
Important disclaimer: The abstract page offers this notice: “Note to users: Uncorrected proofs are Articles in Press that have been copy edited and formatted, but have not been finalized yet. They still need to be proof-read and corrected by the author(s) and the text could still change before final publication.”
Contact: Bill Hathaway – Yale
Source: Yale press release
Image Source: The image is credited to Alan Cowen and is adapted from the Yale press release.
Original Research: Abstract for “Neural portraits of perception: Reconstructing face images from evoked brain activity” by Alan S. Cowen, Marvin M. Chun, and Brice A. Kuhl in Neuroimage. Published online March 17 2014 doi:10.1016/j.neuroimage.2014.03.018 *In Press, Uncorrected Proof*