Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Human-to-Humanoid Real-Time Whole-Body Teleoperation (2403.04436v1)

Published 7 Mar 2024 in cs.RO, cs.AI, cs.LG, cs.SY, and eess.SY

Abstract: We present Human to Humanoid (H2O), a reinforcement learning (RL) based framework that enables real-time whole-body teleoperation of a full-sized humanoid robot with only an RGB camera. To create a large-scale retargeted motion dataset of human movements for humanoid robots, we propose a scalable "sim-to-data" process to filter and pick feasible motions using a privileged motion imitator. Afterwards, we train a robust real-time humanoid motion imitator in simulation using these refined motions and transfer it to the real humanoid robot in a zero-shot manner. We successfully achieve teleoperation of dynamic whole-body motions in real-world scenarios, including walking, back jumping, kicking, turning, waving, pushing, boxing, etc. To the best of our knowledge, this is the first demonstration to achieve learning-based real-time whole-body humanoid teleoperation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (60)
  1. “Teleoperation of humanoid robots: A survey” In IEEE Transactions on Robotics IEEE, 2023
  2. Zipeng Fu, Tony Z Zhao and Chelsea Finn “Mobile aloha: Learning bimanual mobile manipulation with low-cost whole-body teleoperation” In arXiv preprint arXiv:2401.02117, 2024
  3. “Universal Manipulation Interface: In-The-Wild Robot Teaching Without In-The-Wild Robots” In arXiv preprint arXiv:2402.10329, 2024
  4. “Learning fine-grained bimanual manipulation with low-cost hardware” In arXiv preprint arXiv:2304.13705, 2023
  5. “Anthropomorphic movement analysis and synthesis: A survey of methods and applications” In IEEE Transactions on Robotics 32.4 IEEE, 2016, pp. 776–795
  6. Francisco-Javier Montecillo-Puente, Manish Sreenivasa and Jean-Paul Laumond “On real-time whole-body human to humanoid motion transfer”, 2010
  7. “High speed whole body dynamic motion experiment with real time master-slave humanoid robot system” In 2018 IEEE International Conference on Robotics and Automation (ICRA), 2018, pp. 5835–5841 IEEE
  8. “Humanoid dynamic synchronization through whole-body bilateral feedback teleoperation” In IEEE Transactions on Robotics 34.4 IEEE, 2018, pp. 953–965
  9. “Bilateral humanoid teleoperation system using whole-body exoskeleton cockpit TABLIS” In IEEE Robotics and Automation Letters 5.4 IEEE, 2020, pp. 6419–6426
  10. Katsu Yamane, Stuart O Anderson and Jessica K Hodgins “Controlling humanoid robots with human motion data: Experimental validation” In 2010 10th IEEE-RAS International Conference on Humanoid Robots, 2010, pp. 504–510 IEEE
  11. “Slomo: A general system for legged robot motion imitation from casual videos” In IEEE Robotics and Automation Letters IEEE, 2023
  12. “Multi-contact motion retargeting from human to humanoid robot” In 2016 IEEE-RAS 16th international conference on humanoid robots (humanoids), 2016, pp. 1081–1086 IEEE
  13. “Adaptive whole-body manipulation in human-to-humanoid multi-contact motion retargeting” In 2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids), 2017, pp. 446–453 IEEE
  14. “Dynamic locomotion synchronization of bipedal robot and human operator via bilateral feedback teleoperation” In Science Robotics 4.35 American Association for the Advancement of Science, 2019, pp. eaav4282
  15. “Deepmimic: Example-guided deep reinforcement learning of physics-based character skills” In ACM Transactions On Graphics (TOG) 37.4 ACM New York, NY, USA, 2018, pp. 1–14
  16. Jungdam Won, Deepak Gopinath and Jessica Hodgins “A scalable approach to control diverse behaviors for physically simulated characters” In ACM Transactions on Graphics (TOG) 39.4 ACM New York, NY, USA, 2020, pp. 33–1
  17. “Ase: Large-scale reusable adversarial skill embeddings for physically simulated characters” In ACM Transactions On Graphics (TOG) 41.4 ACM New York, NY, USA, 2022, pp. 1–17
  18. “Perpetual Humanoid Control for Real-time Simulated Avatars” In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 10895–10904
  19. “Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control” In arXiv preprint arXiv:2401.16889, 2024
  20. “Learning Humanoid Locomotion with Transformers” In arXiv preprint arXiv:2303.03381, 2023
  21. “Blind bipedal stair traversal via sim-to-real reinforcement learning” In arXiv preprint arXiv:2105.08328, 2021
  22. “Expressive Whole-Body Control for Humanoid Robots” In arXiv preprint arXiv:2402.16796, 2024
  23. “Whole-body geometric retargeting for humanoid robots” In 2019 IEEE-RAS 19th International Conference on Humanoid Robots (Humanoids), 2019, pp. 679–686 IEEE
  24. “A cybernetic avatar system to embody human telepresence for connectivity, exploration, and skill transfer” In International Journal of Social Robotics Springer, 2024, pp. 1–28
  25. “Humanoid Locomotion as Next Token Prediction” In arXiv preprint arXiv:2402.19469, 2024
  26. “AMASS: Archive of motion capture as surface shapes” In Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 5442–5451
  27. “Hybrik: A hybrid analytical-neural inverse kinematics solution for 3d human pose and shape estimation” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 3383–3393
  28. “Deeploco: Dynamic locomotion skills using hierarchical deep reinforcement learning” In ACM Transactions on Graphics (TOG) 36.4 ACM New York, NY, USA, 2017, pp. 1–13
  29. “Unicon: Universal neural controller for physics-based character motion” In arXiv preprint arXiv:2011.15119, 2020
  30. Levi Fussell, Kevin Bergamin and Daniel Holden “Supertrack: Motion tracking for physically simulated characters using supervised learning” In ACM Transactions on Graphics (TOG) 40.6 ACM New York, NY, USA, 2021, pp. 1–13
  31. “Amp: Adversarial motion priors for stylized physics-based character control” In ACM Transactions on Graphics (ToG) 40.4 ACM New York, NY, USA, 2021, pp. 1–20
  32. “Universal Humanoid Motion Representations for Physics-Based Control” In The Twelfth International Conference on Learning Representations, 2024
  33. Alexander Winkler, Jungdam Won and Yuting Ye “QuestSim: Human motion tracking from sparse sensors with simulated avatars” In SIGGRAPH Asia 2022 Conference Papers, 2022, pp. 1–8
  34. “Dynamics-regulated kinematic policy for egocentric pose estimation” In Advances in Neural Information Processing Systems 34, 2021, pp. 25019–25032
  35. Bullet Physics “humanoid urdf in bullet3” Accessed: 2024-03-01, 2023
  36. “Residual force control for agile human behavior imitation and extended motion synthesis” In Advances in Neural Information Processing Systems 33, 2020, pp. 21763–21774
  37. “Robust real-time whole-body motion retargeting from human to humanoid” In 2018 IEEE-RAS 18th International Conference on Humanoid Robots (Humanoids), 2018, pp. 425–432 IEEE
  38. Jonas Koenemann, Felix Burget and Maren Bennewitz “Real-time imitation of human whole-body motions by humanoids” In 2014 IEEE International Conference on Robotics and Automation (ICRA), 2014, pp. 2806–2812 IEEE
  39. “Dancing humanoid robots: Systematic use of osid to compute dynamically consistent movements following a motion capture pattern” In IEEE Robotics & Automation Magazine 22.4 IEEE, 2015, pp. 16–26
  40. “Mixed Reality Teleoperation Assistance for Direct Control of Humanoids” In IEEE Robotics and Automation Letters IEEE, 2024
  41. “Motion retargeting for humanoid robots based on simultaneous morphing parameter identification and motion optimization” In IEEE Transactions on Robotics 33.6 IEEE, 2017, pp. 1343–1357
  42. Kai Hu, Christian Ott and Dongheui Lee “Online human walking imitation in task and joint space based on quadratic programming” In 2014 IEEE International Conference on Robotics and Automation (ICRA), 2014, pp. 3458–3464 IEEE
  43. “Imitate and repurpose: Learning reusable robot movement skills from human and animal behaviors” In arXiv preprint arXiv:2203.17138, 2022
  44. “Humanmimic: Learning natural locomotion and transitions for humanoid robot via wasserstein adversarial imitation” In arXiv preprint arXiv:2309.14225, 2023
  45. “Deep Imitation Learning for Humanoid Loco-manipulation through Human Teleoperation” In 2023 IEEE-RAS 22nd International Conference on Humanoid Robots (Humanoids), 2023, pp. 1–8 IEEE
  46. “iCub3 avatar system: Enabling remote fully immersive embodiment of humanoid robots” In Science Robotics 9.86 American Association for the Advancement of Science, 2024, pp. eadh3834
  47. Jean Chagas Vaz, Dylan Wallace and Paul Y Oh “Humanoid loco-manipulation of pushed carts utilizing virtual reality teleoperation” In ASME International Mechanical Engineering Congress and Exposition 85628, 2021, pp. V07BT07A027 American Society of Mechanical Engineers
  48. “Telexistence and teleoperation for walking humanoid robots” In Intelligent Systems and Applications: Proceedings of the 2019 Intelligent Systems Conference (IntelliSys) Volume 2, 2020, pp. 1106–1121 Springer
  49. Susumu Tachi, Yasuyuki Inoue and Fumihiro Kato “Telesar vi: Telexistence surrogate anthropomorphic robot vi” In International Journal of Humanoid Robotics 17.05 World Scientific, 2020, pp. 2050019
  50. “Alma-articulated locomotion and manipulation for a torque-controllable robot” In 2019 International conference on robotics and automation (ICRA), 2019, pp. 8477–8483 IEEE
  51. “Nimbro wins ana avatar xprize immersive telepresence competition: Human-centric evaluation and lessons learned” In International Journal of Social Robotics Springer, 2023, pp. 1–25
  52. “Proximal Policy Optimization Algorithms” In CoRR abs/1707.06347, 2017 arXiv: http://arxiv.org/abs/1707.06347
  53. “SMPL: A Skinned Multi-Person Linear Model” In ACM Trans. Graphics (Proc. SIGGRAPH Asia) 34.6 ACM, 2015, pp. 248:1–248:16
  54. Diederik P Kingma and Jimmy Ba “Adam: A method for stochastic optimization” In arXiv preprint arXiv:1412.6980, 2014
  55. “Learning to Walk in Minutes Using Massively Parallel Deep Reinforcement Learning”, 2022 arXiv:2109.11978 [cs.RO]
  56. “Learning and Deploying Robust Locomotion Policies with Minimal Dynamics Randomization”, 2023 arXiv:2209.12878 [cs.RO]
  57. “Agile But Safe: Learning Collision-Free High-Speed Legged Locomotion”, 2024 arXiv:2401.17583 [cs.RO]
  58. “Sim-to-real transfer of robotic control with dynamics randomization” In 2018 IEEE international conference on robotics and automation (ICRA), 2018, pp. 3803–3810 IEEE
  59. “Touch and go: Learning from human-collected vision and touch” In arXiv preprint arXiv:2211.12498, 2022
  60. “Language to Action: Towards Interactive Task Learning with Physical Agents.” In IJCAI, 2018, pp. 2–9
Citations (40)

Summary

  • The paper introduces a novel framework that leverages reinforcement learning and a sim-to-data process to retarget human motions for real-time, whole-body humanoid teleoperation.
  • The paper details a two-step retargeting method that adapts the human body model to the humanoid structure and refines motions for feasibility under real-world constraints.
  • The paper demonstrates successful zero-shot transfer from simulation to real-world execution on a Unitree H1 robot, achieving dynamic motion replication with reduced motion data.

Learning Human-to-Humanoid Real-Time Whole-Body Teleoperation

Introduction

Humanoid robots, by virtue of their design, offer unique advantages for tasks that require human-like interaction with the environment. However, controlling these robots to replicate the wide range of human motions in real-time presents a substantial challenge. Conventional model-based approaches to humanoid teleoperation often require simplifications and are heavily reliant on external sensor setups, limiting their applicability in dynamic tasks. Recent advancements in reinforcement learning offer promising solutions, but the application of these techniques to real-world humanoid teleoperation, particularly at the whole-body level, remains largely unexplored.

This paper presents Human to Humanoid (\method), a novel framework that enables the teleoperation of a full-sized humanoid robot using only an RGB camera. The approach is grounded in reinforcement learning, augmented with a "sim-to-data" process that refines a large-scale human motion dataset for feasibility with real-world humanoid constraints. Crucially, this system achieves zero-shot transfer from simulation to real-world application, demonstrating the robot's capability to mimic dynamic human motions such as walking, kicking, and complex gesturing in real time.

Methodology

Retargeting Human Motions

A key innovation in \method is its scalable approach to retargeting human motion for humanoid robots. This process involves adjusting large-scale human motion data to fit the physical constraints and capabilities of a humanoid. The paper introduces a two-step retargeting process, initially adapting the human body model (SMPL) to match the humanoid’s structure, followed by a novel "sim-to-data" method. This method employs a privileged motion imitator to filter out infeasible motions from the retargeted dataset, resulting in a refined set of motions which the real-world humanoid can feasibly execute.

Real-Time Whole-Body Teleoperation Training

The paper articulates a comprehensive training regimen for the humanoid robot, utilizing the retargeted and refined motion dataset. Key to this process is the formulation of an appropriate state space that captures essential motion details while remaining computationally tractable for real-time application. The framework incorporates advanced domain randomization techniques to ensure robustness and generalization of the control policy from simulation to real-world execution.

Evaluation and Results

The \method framework was rigorously tested in both simulated environments and real-world scenarios. In simulation, \method demonstrated superior performance in motion tracking accuracy and success rate over baseline approaches. Notably, the framework exhibited significant resilience to data reduction, maintaining high success rates even when trained on a substantially smaller motion dataset. In real-world tests, \method enabled a full-sized Unitree H1 humanoid robot to replicate a wide variety of dynamic human motions with high fidelity, signifying a substantial leap forward in humanoid teleoperation capabilities.

Implications and Future Directions

The success of \method in enabling advanced humanoid teleoperation has broad implications for the use of humanoid robots in environments that demand human-like dexterity and adaptability. Looking ahead, the paper outlines potential avenues for further research, including improving the representation of motion goals, closing the embodiment gap between humans and robots, and advancements in human-robot interaction to enhance teleoperation efficiency and intuitiveness.

In summary, \method represents a significative advancement in the field of humanoid robotics, offering a robust and scalable solution for real-time whole-body teleoperation using only an RGB camera. This work not only extends the frontier of humanoid robot control but also paves the way for greater human-robot collaboration in complex real-world tasks.

Youtube Logo Streamline Icon: https://streamlinehq.com