News Release

Animal-inspired AI robot learns to navigate unfamiliar terrain

Peer-Reviewed Publication

University of Leeds

Robot adapts gait when tacking real-world terrain

video: 

Robot adapts gait to recover from slips and trips on terrain including muddy grass and a pile of loose timber. 

Credit: Joseph Humphreys, University of Leeds

view more 

Credit: Credit: Joseph Humphreys, University of Leeds

University of Leeds news | Peer-reviewed |  Under embargo until 10am BST Friday, July 11, 2025 

WITH PICS & VIDEOS 

Researchers have developed an Artificial Intelligence (AI) system that enables a four-legged robot to adapt its gait to different, unfamiliar terrain, just like a real animal, in what is believed to be a world first. 

The pioneering technology allows the robot to change the way it moves autonomously, rather than having to be told when and how to alter its stride like the current generation of robots. This advance is seen as a major step towards potentially using legged robots in hazardous settings where humans might be put at risk, such as nuclear decommissioning or search and rescue, where the inability to adapt to the unknown could cost lives. 

For the study, conducted by the University of Leeds and University College London (UCL), the researchers took inspiration from the animal kingdom to teach the robot to navigate terrain that it had never seen before. This included four-legged animals such as dogs, cats and horses, which are adept at adjusting to different landscapes. These animals switch the way they move to save energy, maintain balance, or respond quickly to threats.  

The researchers have created a framework that can teach robots how to transition between trotting, running, bounding and more, just like mammals do in nature. 

Switching gaits when needed 

By embedding within the AI system the same strategies animals use to navigate an unpredictable world, the robot rapidly learns to switch gaits on the fly, in response to the terrain. Thanks to the data processing power of AI, the robot – nicknamed “Clarence”– learned the necessary strategies in just nine hours, considerably faster than the days or weeks most young animals take to confidently cross different surfaces. 

In a paper published today (July 11) in Nature Machine Intelligence, first author Joseph Humphreys, postgraduate researcher in the School of Mechanical Engineering at Leeds, explains how the framework enables the robot to change its stride in accordance with its environment, overcoming a variety of terrains including uneven timber, loose wood chips, and overgrown vegetation, without any alterations to the system itself.   

He said: “Our findings could have a significant impact on the future of legged robot motion control by reducing many of the previous limitations around adaptability.”  

He added: “This deep reinforcement learning framework teaches gait strategies and behaviour inspired by real animals – or ‘bio-inspired’ – such as saving energy, adjusting movements as needed, and gait memory, to achieve highly adaptable and optimal movement, even in environments never previously encountered. 

“All of the training happens in simulation. You train the policy on a computer, then take it and put it on the robot and it is just as proficient as in the training. It’s similar to the Matrix, when Neo's skill in martial arts is downloaded into his brain, but he doesn’t undergo any physical training in the real world. 

“We then tested the robot in the real-world, on surfaces it had never experienced before, and it successfully navigated them all. It was really rewarding to watch it adapt to all the challenges we set and seeing how the animal behaviour we had studied had become almost second nature for it.” 

Deep reinforcement learning agents are often good at learning a specific task but struggle to adapt when the environment changes. Animal brains have built-in structures and information that support learning. Some agents can imitate this kind of learning, but their artificial systems are usually not as advanced or complex. The researchers say they overcame this challenge by instilling their system with natural animal motion strategies. 

They say theirs is the first framework to simultaneously integrate all three critical components of animal locomotion into a reinforcement learning system—namely: gait transition strategies, gait procedural memory, and adaptive motion adjustment—enabling truly versatile, real-world deployment directly from simulation, without needing further adjustment on the physical robot 

In simple terms, the robot doesn’t just learn how to move — it learns how to decide which gait to use, when to switch, and how to adjust it in real time, even on terrain it has never encountered before.  

Professor Zhou, senior author of the study from UCL Computer Science, said: “This research was driven by a fundamental question: what if legged robots could move instinctively the way animals do? Instead of training robots for specific tasks, we wanted to give them the strategic intelligence animals use to adapt their gaits — using principles like balance, coordination, and energy efficiency. 

“By embedding those principles into an AI system, we’ve enabled robots to choose how to move based on real-time conditions, not pre-programmed rules. That means they can navigate unfamiliar environments safely and effectively, even those that they haven’t encountered before. 

“Our long-term vision is to develop embodied AI systems — including humanoid robots — that move, adapt, and interact with the same fluidity and resilience as animals and humans.” 

Real-world applications 

Engineers are increasingly imitating nature — known as biomimicry — to solve complex mobility challenges. The team say their achievement marks a major step forward in making legged robots more adaptable and capable of handling real-world challenges, in hazardous environments or where access is difficult. A robot capable of navigating unfamiliar, complex terrain opens up new possibilities for them to be used in disaster response, planetary exploration, agriculture and infrastructure inspection. 

It also suggests a promising pathway for integrating biological intelligence into robotic systems and conducting more ethical investigations of biomechanics hypotheses; instead of burdening animals with invasive sensors or putting them in danger to study their stability recovery response, robots can be used instead.  

By taking inspiration from factors that make animal movement effective, the researchers were able to develop a framework capable of traversing complex and high-risk terrain despite the robot not using exteroceptive sensors – those being sight, smell and hearing, that help humans in their movements. 

Parallel practice on multiple terrains 

Using deep reinforcement learning – effectively super-powered trial and error – the robot simultaneously practised within hundreds of environments, solving first the challenge of moving with different gaits then choosing the best gait for the terrain, generating the tools to achieve highly adaptable movement.  

To test this acquired adaptability in the real world, the robot was turned loose on real-life surfaces including woodchip, rocks, overgrown roots and loose timber, as well as having its legs repeatedly bashed by a sweeping brush, testing its ability to recover from trips. The team used a programmed route or a joystick – like those used in video games – to direct the robot. 

Perhaps surprisingly, the robot was not exposed to any rough terrain during training, highlighting the system's ability to adapt and demonstrating that these skills have become instinctive for the robot. 

The study, part-funded by the Royal Society and the Advanced Research and Invention Agency (ARIA), focused on enabling robust everyday movement. In future work, the team hope to add more dynamic skills, such as long-distance jumping, climbing, and navigating steep or vertical terrains. 

Although the framework has so far only been tested on a single dog-sized quadruped robot, the underlying principles are broadly applicable. The same bio-inspired metrics can be used across a wide range of four-legged robots, regardless of size or weight, as long as they share a similar morphology. 

Further Information 

The paper ‘Learning to Adapt through Bio-Inspired Gait Strategies for Versatile Quadruped Locomotion’ is published in Nature Machine Intelligence on Friday July 11, 2025. 

DOI: 10.1038/s42256-025-01065-z 

For media inquiries and interview requests please contact Deb Newman via d.newman@leeds.ac.uk and copy in pressoffice@leeds.ac.uk


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.