Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 102 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 30 tok/s
GPT-5 High 27 tok/s Pro
GPT-4o 110 tok/s
GPT OSS 120B 475 tok/s Pro
Kimi K2 203 tok/s Pro
2000 character limit reached

Learning Humanoid Locomotion with Perceptive Internal Model (2411.14386v1)

Published 21 Nov 2024 in cs.RO

Abstract: In contrast to quadruped robots that can navigate diverse terrains using a "blind" policy, humanoid robots require accurate perception for stable locomotion due to their high degrees of freedom and inherently unstable morphology. However, incorporating perceptual signals often introduces additional disturbances to the system, potentially reducing its robustness, generalizability, and efficiency. This paper presents the Perceptive Internal Model (PIM), which relies on onboard, continuously updated elevation maps centered around the robot to perceive its surroundings. We train the policy using ground-truth obstacle heights surrounding the robot in simulation, optimizing it based on the Hybrid Internal Model (HIM), and perform inference with heights sampled from the constructed elevation map. Unlike previous methods that directly encode depth maps or raw point clouds, our approach allows the robot to perceive the terrain beneath its feet clearly and is less affected by camera movement or noise. Furthermore, since depth map rendering is not required in simulation, our method introduces minimal additional computational costs and can train the policy in 3 hours on an RTX 4090 GPU. We verify the effectiveness of our method across various humanoid robots, various indoor and outdoor terrains, stairs, and various sensor configurations. Our method can enable a humanoid robot to continuously climb stairs and has the potential to serve as a foundational algorithm for the development of future humanoid control methods.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper presents the Perceptive Internal Model that integrates LiDAR-based elevation maps to enhance humanoid locomotion by addressing inherent instability and complex terrain challenges.
  • It employs a Hybrid Internal Model trained in simulation with ground-truth obstacle heights, achieving robust navigation across diverse indoor and outdoor environments.
  • The research demonstrates impressive computational efficiency, completing training in three hours on an RTX 4090 GPU and setting a benchmark for sim-to-real transitions in robotics.

Learning Humanoid Locomotion with Perceptive Internal Model

The paper "Learning Humanoid Locomotion with Perceptive Internal Model" addresses the complex problem of achieving stable and dynamic locomotion in humanoid robots, which possess high degrees of freedom and inherently unstable morphologies. Unlike quadruped robots that can rely on a "blind" policy to traverse various terrains, humanoid robots necessitate precise perception for stable and adaptive movement. The introduction of perceptual signals, however, can introduce additional disturbances and complexities, potentially reducing system robustness and efficiency.

This research proposes the Perceptive Internal Model (PIM), a novel approach that uses onboard, continuously updated elevation maps for terrain perception around the robot. The PIM, trained in a simulated environment using ground-truth obstacle heights, integrates perceptive information to optimize robot locomotion through a simulated internal model framework known as the Hybrid Internal Model (HIM). This strategy contrasts with previous efforts that encode raw point clouds or depth maps, which are more susceptible to camera noise and movement disturbances. The methodology employed by PIM involves training in simulation with ground-truth mappings and uses LiDAR-based elevation maps for perception during real-world inference, which allows for substantial computational efficiency.

The results presented in the paper are notable in terms of computational efficiency, highlighting that the training process can be completed within three hours on an RTX 4090 GPU. This efficiency is achieved without the need for depth map rendering in simulation, thereby avoiding additional computational costs. The outcomes demonstrate the method's robustness across a variety of humanoid robots and terrains, including both indoor and outdoor environments, and validate its ability to enable continuous stair climbing at a high success rate.

The implications of this research are significant for the field of humanoid robotics. By effectively integrating perceptive information into locomotion policy, this work lays the groundwork for future developments in humanoid control methods, offering a more reliable foundation for robots to handle complex and dynamic environments. The PIM has the potential to ameliorate existing challenges in perceptive legged locomotion, providing a framework that can support the development of more agile and adaptive humanoid robots.

Future developments stemming from this work may lead to enhancements in humanoid robotic navigation and task execution in various real-world settings. The research opens avenues for further exploration into sim2real transitions, where the fidelity of simulation to real-world performance can be fine-tuned and optimized beyond the robust results already achieved in this paper. Additionally, the exploitation of more complex sensor fusion strategies and machine perception technologies can be anticipated as a natural extension of the foundational methodologies introduced here. This research offers profound insights and concrete methods that can be adopted and adapted by robotics researchers aiming to push the boundaries of humanoid robot capabilities.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.