Home Mind & Brain New Research Sheds Light on How Human Vision Perceives Scale

New Research Sheds Light on How Human Vision Perceives Scale

Published: Last updated:
Reading Time: 2 minutes

Researchers from Aston University and the University of York have discovered new insights into how the human brain perceives the external world.

The study explored the computational mechanisms the human brain uses to perceive the size of objects in the world around us.

The research, led by Professor Tim Meese, in the School of Optometry at Aston University and Dr Daniel Baker in the Department of Psychology at the University of York, tells us more about how our visual system can exploit ‘defocus blur’ to infer perceptual scale, but that it does so crudely.

It is well known that to derive object size from retinal image size; our visual system needs to estimate the distance to the object. The retinal image contains many pictorial cues, such as linear perspective, which help the system derive the relative size of objects. However, the system needs to know about spatial scale to derive absolute size.

By taking account of defocus blur, like the blurry parts of an image outside the depth of focus of a camera, the visual system can achieve this. Others have worked out the maths behind this, but the study asked: does human vision exploit this maths?

The research team presented participants with photographic pairs of full-scale railway scenes subject to various artificial blur treatments and small-scale models of railway scenes taken with long exposure and a small aperture to diminish defocus blur. The task was to detect which photograph was the real full-scale scene in each pair.

When the artificial blur was appropriately oriented with the ground plane (the horizontal plane representing the ground on which the viewer is standing) in the full-scale scenes, participants were fooled and believed the small models to be the full-scale scenes. Remarkably, this did not require the application of realistic gradients of blur. Simple uniform bands of blur at the top and bottom of the photographs achieved almost equivalent miniaturisation effects.

Tim Meese, professor of vision science at Aston University, said: “Our results indicate that human vision can exploit defocus blur to infer perceptual scale but do this crudely – more a heuristic than a metrical analysis. Overall, our findings provide new insights into the computational mechanisms used by the human brain in perceptual judgments about the relationship between ourselves and the external world.”

Daniel Baker, senior lecturer in psychology at the University of York, said: “These findings demonstrate that our perception of size is not perfect and can be influenced by other properties of a scene. It also highlights the remarkable adaptability of the visual system. This might have relevance for understanding the computational principles underlying our perception of the world. For example, when judging the size and distance of hazards when driving.”

© Copyright 2014–2034 Psychreg Ltd