Knowledge about the real-world size of an object, not its visual size, guides one’s attention, according to a new GW study.
A stop sign at an intersection appears to be larger than a parked car across the street. Is our brain playing tricks on us?
In a recently published study in the journal Nature Human Behavior, Professor of Cognitive Neuroscience Sarah Shomstein uncovered important insights into how the human brain processes information and allocates attention. The study, “Attention Scales According to Inferred Real-World Object Size,” shows people pay attention to objects based on their real-world size, rather than how they are perceived by the eye.
“Since a person can only pay attention to a limited amount of information at a time, our brain uses object size to determine how much attention to allocate to that object,” Shomstein said. “However, the way our eyes perceive an object can be different from its actual size, such as a car appearing large when it is close and small when it is far. Our study has shown for the first time that the brain adjusts attention based on our knowledge of an object’s size, not how our eyes view it.”
The findings could have significant impact on many everyday tasks, from a radiologist looking for tumors of various sizes to an airport security agent tasked with searching for threats hidden in luggage. The research, which was funded by a grant from the National Science Foundation, suggests attention allocation can be trained, which could lead to improvements in efficiently performing such tasks.
To determine whether the human brain differentially allocates attention depending on an object’s size, Shomstein and her team showed participants several images of small and large everyday objects such as a domino or billiards table, but presented them at an identical fixed size. The team measured how long it took participants to identify probe targets that were embedded within object images. They found that responses were quicker across known smaller real-world sized objects compared to larger ones, even though the space occupied by both objects on the eye was the same.
“If objects are of identical size on your eye, but you know that one of them is smaller—such as a domino nearby versus a pool table far away—you allocate more attention to the smaller item,” Shomstein explained. “Your knowledge of the object’s size overrides its physical size thereby adjusting how much attention your brain allocates to the objects.”
The research team also showed participants images of everyday objects on a screen and asked them to rate the size of each object on a scale of one to six, one being very small and six being very large. Findings showed a direct correlation between how participants rated the size of each object and how long it took them to respond to target stimuli within the image.
“Your own personal ratings determine how efficient you are going to be at attending to that object,” said co-author Andrew J. Collegio, MPhil ’18, PhD ’18, a researcher in Shomstein’s Attention and Cognition Lab. “If you think the pool table is really large, then your attention is going to be less focused.”
Shomstein and Collegio hope these findings will help future researchers predict how effectively people process particular objects as they pay attention to the world around them.