A rope coiled on dusty trail may trigger a frightened jump by a hiker who recently stepped on a snake. Now a new study better explains how a one-time visual experience can shape perceptions afterward.
Led by neuroscientists from NYU School of Medicine and published online in eLife, the study argues that humans recognize what they are looking at by combining current sensory stimuli with comparisons to images stored in memory.
“Our findings provide important new details about how experience alters the content-specific activity in brain regions not previously linked to the representation of images by nerve cell networks,” says senior study author Biyu J. He, PhD, assistant professor in the departments of Neurology, Radiology, and Neuroscience and Physiology.
“The work also supports the theory that what we recognize is influenced more by past experiences than by newly arriving sensory input from the eyes,” says Dr. He, who is also part of the Neuroscience Institute at NYU Langone Health.
She says this idea becomes more important as evidence mounts that hallucinations suffered by patients with post-traumatic stress disorder or schizophrenia occur when stored representations of past images overwhelm what they are looking at presently.
Glimpse of a Tiger
A key question in neurology is about how the brain perceives, for instance, that a tiger is nearby based on a glimpse of orange amid the jungle leaves. If the brains of our ancestors matched this incomplete picture with previous danger, they would be more likely to hide, survive, and have descendants. Thus, the modern brain finishes perception puzzles without all the pieces.
Most past vision research, however, has been based on experiments wherein clear images were shown to subjects in perfect lighting, says Dr. He. The current study instead analyzed visual perception as subjects looked at black-and-white images degraded until they were difficult to recognize.
Nineteen subjects were shown 33 such obscured “Mooney images”—17 of animals and 16 manmade objects—in a particular order. They viewed each obscured image six times, then a corresponding clear version once to achieve recognition, and then blurred images again six times after. Following the presentation of each blurred image, subjects were asked if they could name the object shown.
As the subjects sought to recognize images, the researchers “took pictures” of their brains every two seconds using functional magnetic resonance imaging (fMRI). The technology lights up with increased blood flow, which is known to happen as brain cells are turned on during a specific task. The team’s 7 Tesla scanner offered a more than three-fold improvement in resolution over past studies using standard 3 Tesla scanners, for extremely precise fMRI-based measurement of vision-related nerve circuit activity patterns.
After seeing the clear version of each image, the study subjects were more than twice as likely to recognize what they were looking at when again shown the obscured version as they were of recognizing it before seeing the clear version. They had been “forced” to use a stored representation of clear images, called priors, to better recognize related, blurred versions, says Dr. He.
The authors then used mathematical tricks to create a two-dimensional map that measured, not nerve cell activity in each tiny section of the brain as it perceived images, but instead similarities between nerve network activity patterns in different brain regions. Nerve cell networks in the brain that represented images more similarly landed closer to each other on the map.
This approach revealed the existence of a stable system of brain organization that processed each image in the same steps, and regardless of whether clear or blurry, the authors say. Early, simpler brain circuits in the visual cortex that determine edge, shape, and color clustered on one end of the map, and more complex, “higher-order” circuits known to mix past and present information to plan actions at the opposite end.
These higher-order circuits included two brain networks, the default-mode network (DMN) and frontoparietal network (FPN), both linked by past studies to executing complex tasks such as planning actions, but not to visual, perceptual processing. Rather than remaining stable in the face of all images, the similarity patterns in these two networks shifted as brains went from processing unrecognized, blurry images to effortlessly recognizing the same images after seeing a clear version. After previously seeing a clear version (disambiguation), neural activity patterns corresponding to each blurred image in the two networks became more distinct from the others, and more like the clear version in each case.
Strikingly, the clear image-induced shift of neural representation towards perceptual prior was much more pronounced in brain regions with higher, more complex functions than in the early, simple visual processing networks. This further suggests that more of the information shaping current perceptions comes from what people have experienced before.
Along with Dr. He, the study authors were Carlos González-García and Alexis Baria of the National Institute of Neurological Disorders and Stroke at the National Institutes of Health, Bethesda, Maryland, as well as Matthew Flounders and Raymond Chang of the Neuroscience Institute at NYU School of Medicine. González-García is currently a postdoctoral fellow in the Department of Experimental Psychology at Ghent University in Belgium.
This research was supported by the Intramural Research Program of the National Institutes of Health/National Institute of Neurological Disorders and Stroke; NYU Langone Health; Leon Levy Foundation; Klingenstein-Simons Fellowship; and the Department of State Fulbright Program. The experiment was approved by the Institutional Review Board of the National Institute of Neurological Disorders and Stroke. All subjects provided written, informed consent.