Nav: Home

Visual biases near the hands help us perform specific actions

December 05, 2016

Using your hands to perform tasks in specific ways can change the way you see things near your hands, according to new research published in Psychological Science, a journal of the Association for Psychological Science. The research shows that learning to grasp an object with the backs of the hands made participants more sensitive to motion near their hands, while learning a to pick up an object with their fingertips enhanced participants' perception of spatial detail near their hands.

"These results support the idea that vision is, in a sense, tuned for action," says psychological scientist Laura Thomas of North Dakota State University, author of the study. "This evidence suggests that people experience flexible visual biases when viewing information within reach that may aid the actions they are prepared to take."

Previous studies have indicated that our visual biases - the visual information we're attuned to at a given time - may adapt to the immediate context. When participants in one study adopted a power-grasp posture, they showed greater sensitivity to motion-related information that would allow them to make a quick and forceful grab. When the participants adopted a precision-grasp posture, they were more sensitive to detailed spatial information that would help them manipulate something small.

These findings suggest that the visual system prioritizes processing information that is relevant to the actions we are likely to take, allowing us to be more effective power or precision graspers, depending on the situation.

Thomas wondered whether a short-term training experience might be sufficient to induce these action-oriented visual biases. To find out, she conducted two experiments in which participants learned to use their hands in new ways to pick up objects.

In the first experiment, 60 student participants completed two visual perception tasks, as their hands either rested on the display or in their laps. In one task, they saw a group of moving dots--some of which moved in the same direction--and had to indicate the direction in which the dots appeared to be moving. In the other task, they saw another group of stationary dots--some of which were arranged in a particular spatial pattern--and had to say whether the dots appeared in a radial or concentric spatial pattern.

The participants then practiced a new type of power grasp, using the backs of their hands to pick up and move a plunger.

After this training, the participants completed the two visual perception tasks again.

In the second experiment, another group of 60 student participants followed the same procedure - this time, they learned a new type of precision grasp, using the tips of their little fingers to pick up and move one bean.

The results of the two experiments showed that potential action seemed to drive the way the participants processed visual information, but only when that information was near their hands.

Participants were better at identifying the direction in which the dots were shifting after learning the back-of-the-hands power grasp, but only when their hands were near the display.

And they were more accurate in identifying the spatial pattern of dots after learning the finger-tips precision grasp - again, only when their hands were near the display.

Together, these findings suggest that our visual system flexibly adapts to recent experiences, allowing us to take action in effective and appropriate ways.

Thomas is planning further research to examine this phenomenon, investigating whether the kinds of visual biases she has observed in lab-based tasks also emerge when people look at and interact with everyday objects.

"The more we learn about how visual processing biases near the hands operate, the more potential we have to make recommendations about how to best present information on handheld devices, like smartphones and tablets, to meet users' information processing goals," she says.
This material is based on work supported by the National Science Foundation under Grant BCS 1556336 and by a Google Faculty Research Award.

All data have been made publicly available via the Open Science Framework and can be accessed at The complete Open Practices Disclosure for this article can be found at This article has received the badge for Open Data. More information about the Open Practices badges can be found at and

For more information about this study, please contact: Laura E. Thomas at

The article abstract is available online:

The APS journal Psychological Science is the highest ranked empirical journal in psychology. For a copy of the article "Action Experience Drives Visual-Processing Biases Near the Hands" and access to other Psychological Science research findings, please contact Anna Mikulak at 202-293-9300 or

Association for Psychological Science

Related Learning Articles:

When learning on your own is not enough
We make decisions based on not only our own learning experience, but also learning from others.
Learning more about particle collisions with machine learning
A team of Argonne scientists has devised a machine learning algorithm that calculates, with low computational time, how the ATLAS detector in the Large Hadron Collider would respond to the ten times more data expected with a planned upgrade in 2027.
Getting kids moving, and learning
Children are set to move more, improve their skills, and come up with their own creative tennis games with the launch of HomeCourtTennis, a new initiative to assist teachers and coaches with keeping kids active while at home.
How expectations influence learning
During learning, the brain is a prediction engine that continually makes theories about our environment and accurately registers whether an assumption is true or not.
Technology in higher education: learning with it instead of from it
Technology has shifted the way that professors teach students in higher education.
Learning is optimized when we fail 15% of the time
If you're always scoring 100%, you're probably not learning anything new.
School spending cuts triggered by great recession linked to sizable learning losses for learning losses for students in hardest hit areas
Substantial school spending cuts triggered by the Great Recession were associated with sizable losses in academic achievement for students living in counties most affected by the economic downturn, according to a new study published today in AERA Open, a peer-reviewed journal of the American Educational Research Association.
Lessons in learning
A new Harvard study shows that, though students felt like they learned more from traditional lectures, they actually learned more when taking part in active learning classrooms.
Learning to look
A team led by JGI scientists has overhauled the perception of inovirus diversity.
Sleep readies synapses for learning
Synapses in the hippocampus are larger and stronger after sleep deprivation, according to new research in mice published in JNeurosci.
More Learning News and Learning Current Events

Trending Science News

Current Coronavirus (COVID-19) News

Top Science Podcasts

We have hand picked the top science podcasts of 2020.
Now Playing: TED Radio Hour

Warped Reality
False information on the internet makes it harder and harder to know what's true, and the consequences have been devastating. This hour, TED speakers explore ideas around technology and deception. Guests include law professor Danielle Citron, journalist Andrew Marantz, and computer scientist Joy Buolamwini.
Now Playing: Science for the People

#576 Science Communication in Creative Places
When you think of science communication, you might think of TED talks or museum talks or video talks, or... people giving lectures. It's a lot of people talking. But there's more to sci comm than that. This week host Bethany Brookshire talks to three people who have looked at science communication in places you might not expect it. We'll speak with Mauna Dasari, a graduate student at Notre Dame, about making mammals into a March Madness match. We'll talk with Sarah Garner, director of the Pathologists Assistant Program at Tulane University School of Medicine, who takes pathology instruction out of...
Now Playing: Radiolab

What If?
There's plenty of speculation about what Donald Trump might do in the wake of the election. Would he dispute the results if he loses? Would he simply refuse to leave office, or even try to use the military to maintain control? Last summer, Rosa Brooks got together a team of experts and political operatives from both sides of the aisle to ask a slightly different question. Rather than arguing about whether he'd do those things, they dug into what exactly would happen if he did. Part war game part choose your own adventure, Rosa's Transition Integrity Project doesn't give us any predictions, and it isn't a referendum on Trump. Instead, it's a deeply illuminating stress test on our laws, our institutions, and on the commitment to democracy written into the constitution. This episode was reported by Bethel Habte, with help from Tracie Hunte, and produced by Bethel Habte. Jeremy Bloom provided original music. Support Radiolab by becoming a member today at     You can read The Transition Integrity Project's report here.