The robot used trial and error to learn what kind of grip would be successful. During testing, the hand was able to successfully grasp 11 of 14 objects. (Image: Bio-Inspired Robotics Laboratory, University of Cambridge)

Grasping objects of different sizes, shapes and textures is a problem that is easy for a human, but challenging for a robot. Researchers from the University of Cambridge designed a soft, 3D-printed robotic hand that cannot independently move its fingers but can still carry out a range of complex movements.

The robot hand was trained to grasp different objects and was able to predict whether it would drop them by using the information provided by sensors placed on its ’skin’.

This type of passive movement makes the robot far easier to control and far more energy-efficient than robots with fully motorized fingers. The researchers say their adaptable design could be used in the development of low-cost robots that are capable of more natural movement and can learn to grasp a wide range of objects. The results are reported in the journal Advanced Intelligent Systems.

In the natural world, movement results from the interplay between the brain and the body. This enables people and animals to move in complex ways without expending unnecessary amounts of energy. Over the past several years, soft components have begun to be integrated into robotics design thanks to advances in 3D printing techniques, which have allowed researchers to add complexity to simple, energy-efficient systems.

In Professor Fumiya Iida’s Bio-Inspired Robotics Laboratory in Cambridge’s Department of Engineering, researchers have been developing potential solutions to both problems: a robot hand that can grasp a variety of objects with the correct amount of pressure while using a minimal amount of energy.

“In earlier experiments, our lab has shown that it’s possible to get a significant range of motion in a robot hand just by moving the wrist,” said co-author Dr Thomas George-Thuruthel, who is now based at University College London (UCL) East. “We wanted to see whether a robot hand based on passive movement could not only grasp objects, but would be able to predict whether it was going to drop the objects or not, and adapt accordingly.”

The researchers used a 3D-printed anthropomorphic hand implanted with tactile sensors, so that the hand could sense what it was touching. The hand was only capable of passive, wrist-based movement.

The robot hand was trained to grasp different objects and was able to predict whether it would drop them by using the information provided by sensors placed on its skin. (Image: Bio-Inspired Robotics Laboratory, University of Cambridge)

“This kind of hand has a bit of springiness to it: It can pick things up by itself without any actuation of the fingers,” said first author Dr Kieran Gilday, who is now based at EPFL in Lausanne, Switzerland. “The tactile sensors give the robot a sense of how well the grip is going, so it knows when it’s starting to slip. This helps it to predict when things will fail.”

The robot used trial and error to learn what kind of grip would be successful. After finishing the training with the balls, it then attempted to grasp different objects including a peach, a computer mouse, and a roll of bubble wrap. In these tests, the hand was able to successfully grasp 11 of 14 objects.

“The big advantage of this design is the range of motion we can get without using any actuators,” said Iida. “We want to simplify the hand as much as possible. We can get lots of good information and a high degree of control without any actuators, so that when we do add them, we’ll get more complex behavior in a more efficient package.”

A fully actuated robotic hand, in addition to the amount of energy it requires, is also a complex control problem. The passive design of the Cambridge-designed hand, using a small number of sensors, is easier to control, provides a wide range of motion, and streamlines the learning process.

For more information, contact Sarah Collins at This email address is being protected from spambots. You need JavaScript enabled to view it.; +012-237-65542.