Summary: Perception in virtual reality is more strongly influenced by expectations than visual information, researchers report.
Source: University of Western Ontario
A new study by Western neuroscientists suggests that, unlike true reality, perception in virtual reality is more strongly influenced by our expectations than the visual information before our eyes.
The researchers point to the challenge of online shopping, where customers sometimes mis-estimate the size of a product based on their expectations, discovering for example that a sweater purchased online is indeed lovely but sized for a doll not an adult.
This happens in part because the physical cues to size that are present when seeing an item in a store are typically eliminated when viewing photos online. Without seeing the physical object, customers base their expectations of familiar size on prior experience. Since most sweaters are sized for people, not dolls, the visual system assumes that an unfamiliar sweater is, too.
The advent of virtual reality offers new opportunities for applications like online shopping and also for research on visual perception. But researchers wanted to understand if users of virtual reality perceive size as accurately as they do in the real world.
A research team, led by Canada Research Chair in Immersive Neuroscience Jody Culham, presented study participants with a variety of familiar objects like dice and sports balls in virtual reality and asked them to estimate the object sizes. The trick? Objects were presented not only at their typical ‘familiar’ sizes, but also at unusual sizes (e.g., die-sized Rubik’s cubes).
The researchers found that participants consistently perceived the virtual objects at the size they expected, rather than the actual presented size. This effect was much stronger in virtual reality than for real objects.
“While virtual reality is a useful research tool with many real-world applications, we cannot assume it is always an accurate proxy for the real world,” said Culham, a psychology professor and senior author on the study.
“It is promising to see advances in virtual reality and its applications, but there is still a lot we don’t understand about how we process information in virtual environments. If we need to rely heavily on past experiences to judge the size of objects in virtual reality, this suggests other visual cues to size may be less reliable than in the real world.”
Yet, the results of this study also have some promising implications.
“If we know that familiar objects can serve as strong size cues in virtual reality, we can use this information to our advantage,” said Anna Rzepka, a former student in the Culham Lab and co-first author on the study.
“Think about viewing an item in a scene where accurate size perception is crucial, such as when removing a tumor using image-guided surgery. Adding other familiar objects to the virtual scene could improve perception of the tumor’s size and location, leading to better outcomes.”
About this virtual reality and perception research news
Author: Jeff Renaud
Source: University of Western Ontario
Contact: Jeff Renaud – University of Western Ontario
Image: The image is in the public domain
Original Research: Open access.
“Familiar size affects perception differently in virtual reality and the real world” by Anna M. Rzepka et al. Philosophical Transactions of the Royal Society B: Biological Sciences
Familiar size affects perception differently in virtual reality and the real world
The promise of virtual reality (VR) as a tool for perceptual and cognitive research rests on the assumption that perception in virtual environments generalizes to the real world. Here, we conducted two experiments to compare size and distance perception between VR and physical reality (Maltz et al. 2021 J. Vis.21, 1–18).
In experiment 1, we used VR to present dice and Rubik’s cubes at their typical sizes or reversed sizes at distances that maintained a constant visual angle. After viewing the stimuli binocularly (to provide vergence and disparity information) or monocularly, participants manually estimated perceived size and distance.
Unlike physical reality, where participants relied less on familiar size and more on presented size during binocular versus monocular viewing, in VR participants relied heavily on familiar size regardless of the availability of binocular cues.
In experiment 2, we demonstrated that the effects in VR generalized to other stimuli and to a higher quality VR headset. These results suggest that the use of binocular cues and familiar size differs substantially between virtual and physical reality.
A deeper understanding of perceptual differences is necessary before assuming that research outcomes from VR will generalize to the real world.