We present STEPP for traversability estimation and show its working method and real world deployment.
We present STEPP for traversability estimation and show its working method and real world deployment.
Understanding the traversability of terrain is essential for autonomous robot navigation, particularly in unstructured environments such as natural landscapes. While traditional methods, such as occupancy mapping provide a basic framework, they often fail to account for the complex mobility capabilities of some platforms such as legged robots. In this work, we propose a method to estimate terrain traversability by learning from demonstrations of humans walking. Our approach leverages dense, pixel-wise feature embeddings generated using the DINOv2 vision Transformer model, which are processed through an encoder-decoder MLP architecture to analyze terrain segments. The averaged feature vectors, extracted from masked regions of interest, are used to train the model in a reconstruction-based framework. By minimizing reconstruction loss, the network distinguishes between familiar terrain, with low reconstruction error, and unfamiliar or hazardous terrain with higher reconstruction error. This approach facilitates anomaly detection, enabling a legged robot to navigate more effectively through challenging terrain. We run real world experiments on the ANYmal legged robot both indoor and outdoor to prove our proposed method.
Avoiding trees, but navigating medium tall grass.
Navigating around tree and barrier while walking through tall grass.
Navigating through a maze like environment.
Navigating over mats to show traversability for more than just color.
Straight forest road.
Forest road that splits.
Forest trail.
(coming soon)