Someday, when quakes, fires, and floods strike, the first responders might be packs of robotic rescue dogs rushing in to help stranded souls. These battery-powered quadrupeds would use computer vision to size up obstacles and employ doglike agility skills to get past them.
Toward that goal, AI researchers at Stanford University and Shanghai Qi Zhi Institute say they have developed a new vision-based algorithm that helps robodogs scale high objects, leap across gaps, crawl under thresholds, squeeze through crevices, and then bolt to the next challenge. The algorithm represents the brains of the robodog.
“The autonomy and range of complex skills that our quadruped robot learned is quite impressive,” said Chelsea Finn, assistant professor of computer science and senior author of a new peer-reviewed paper announcing the teams’ approach. “And we have created it using low-cost, off-the-shelf robots — actually, two different off-the-shelf robots.”
The key advance, the authors say, is that their robodog is autonomous, that is, it is able to size up physical challenges and imagine, then execute, a broad range of agility skills based simply on the obstacles it sees.
“What we’re doing is combining both perception and control, using images from a depth camera mounted on the robot and machine learning to process all those inputs and move the legs in order to get over, under, and around obstacles,” said Zipeng Fu, a doctoral candidate in Finn’s lab and first author of the study, along with Ziwen Zhuang of Shanghai Qi Zhi Institute.
Theirs is not the first robodog to demonstrate such agility — a class of athletics known as “parkour” — but it is first to combine self-sufficiency with a broad array of skills.
“Our robots have both vision and autonomy, the athletic intelligence to size up a challenge and to self-select and execute parkour skills based on the demands of the moment,” Fu said.
Existing learning methods are often based on complex reward systems that must be fine-tuned to specific physical obstacles. Accordingly, they don’t scale well to new or unfamiliar environments. Other approaches learn using real-world data to imitate agility skills of animals. Those robodogs lack a broad skill set and don’t have the new robodogs’ vision capabilities. And both existing methods are computationally slow.
This is the first open-source application to accomplish these goals with a simple reward system using no real-world reference data, according to the study.
To succeed, they first synthesized and honed the algorithm using a computer model, then transferred it to two real-world robodogs. Next, in a process called reinforcement learning, the robots attempted to move forward in any way they saw fit and got rewarded based on how well they did. This is how the algorithm eventually learns the best way to approach a new challenge.
In practice, most existing reinforcement learning reward systems involve too many variables to be effective, slowing computational performance. This is what makes the streamlined reward process for robodog parkour exceptional and surprisingly straightforward.
“It’s actually fairly simple,” Finn said. “We based it mostly on how far forward the robot is moving and the amount of effort it has applied to do it. Eventually, the robot learns more complex motor skills that allow it to get ahead.”
The team performed extensive experiments using real-world robodogs to demonstrate their new agility approach in especially challenging environments using only those robodogs’ off-the-shelf computers, visual sensors, and power systems.
The new-and-improved robodogs were able to climb obstacles more than one-and-a-half times their height, leap gaps greater than one-and-a-half times their length, crawl beneath barriers three-quarters of their height, and tilt themselves in order to squeeze through a slit thinner than their width.
Next, the team hopes to leverage advances in 3D vision and graphics to add real-world data to its simulated environments to bring a new level of real-world autonomy to their algorithm.