Expressive Whole-Body Control for Humanoid Robots (2402.16796v2)
Abstract: Can we enable humanoid robots to generate rich, diverse, and expressive motions in the real world? We propose to learn a whole-body control policy on a human-sized robot to mimic human motions as realistic as possible. To train such a policy, we leverage the large-scale human motion capture data from the graphics community in a Reinforcement Learning framework. However, directly performing imitation learning with the motion capture dataset would not work on the real humanoid robot, given the large gap in degrees of freedom and physical capabilities. Our method Expressive Whole-Body Control (Exbody) tackles this problem by encouraging the upper humanoid body to imitate a reference motion, while relaxing the imitation constraint on its two legs and only requiring them to follow a given velocity robustly. With training in simulation and Sim2Real transfer, our policy can control a humanoid robot to walk in different styles, shake hands with humans, and even dance with a human in the real world. We conduct extensive studies and comparisons on diverse motions in both simulation and the real world to show the effectiveness of our approach.
- Legged locomotion in challenging terrains using egocentric vision. In Conference on Robot Learning, pages 403–415. PMLR, 2023.
- Legs as manipulator: Pushing quadrupedal agility beyond locomotion. In 2023 IEEE International Conference on Robotics and Automation (ICRA), 2023a.
- Extreme parkour with legged robots. arXiv preprint arXiv:2309.14341, 2023b.
- The mit humanoid robot: Design, motion planning, and control for acrobatic behaviors. In 2020 IEEE-RAS 20th International Conference on Humanoid Robots (Humanoids), pages 1–8. IEEE, 2021.
- Sim-to-real learning for humanoid box loco-manipulation. arXiv preprint arXiv:2310.03191, 2023.
- Whole body humanoid control from human motion descriptors. In 2008 IEEE International Conference on Robotics and Automation, pages 2677–2684. IEEE, 2008.
- Learning vision-based bipedal locomotion for challenging terrain. arXiv preprint arXiv:2309.14594, 2023.
- Adversarial motion priors make good substitutes for complex reward functions. 2022 ieee. In International Conference on Intelligent Robots and Systems (IROS), volume 2, 2022a.
- Adversarial motion priors make good substitutes for complex reward functions. In 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 25–32. IEEE, 2022b.
- Learning deep sensorimotor policies for vision-based autonomous drone racing, 2022.
- Minimizing energy consumption leads to the emergence of gaits in legged robots. Conference on Robot Learning (CoRL), 2021.
- Deep whole-body control: learning a unified policy for manipulation and locomotion. In Conference on Robot Learning, pages 138–149. PMLR, 2023.
- Opt-mimic: Imitation of optimized trajectories for dynamic quadruped behaviors. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 5092–5098. IEEE, 2023.
- Mabel, a new robotic bipedal walker and runner. In 2009 American Control Conference, pages 2030–2036. IEEE, 2009.
- Generating diverse and natural 3d human motions from text. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5152–5161, June 2022.
- Synthesizing physical character-scene interactions. 2023. doi: 10.1145/3588432.3591525. URL https://doi.org/10.1145/3588432.3591525.
- The development of honda humanoid robot. In Proceedings. 1998 IEEE international conference on robotics and automation (Cat. No. 98CH36146), volume 2, pages 1321–1326. IEEE, 1998.
- Anymal-a highly mobile and dynamic quadrupedal robot. In IROS, 2016.
- Efficient multitask learning with an embodied predictive model for door opening and entry with whole-body control. Science Robotics, 7(65):eaax8177, 2022.
- Learning whole-body manipulation for quadrupedal robot. arXiv preprint arXiv:2308.16820, 2023.
- Hierarchical reinforcement learning for precise soccer shooting skills using a quadrupedal robot. In 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 1479–1486. IEEE, 2022.
- Dribblebot: Dynamic legged manipulation in the wild. arXiv preprint arXiv:2304.01159, 2023.
- The 3d linear inverted pendulum mode: A simple modeling for a biped walking pattern generation. In Proceedings 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems. Expanding the Societal Role of Robotics in the the Next Millennium (Cat. No. 01CH37180), volume 1, pages 239–246. IEEE, 2001.
- Ichiro Kato. Development of wabot 1. Biomechanism, 2:173–214, 1973.
- Rma: Rapid motor adaptation for legged robots. arXiv preprint arXiv:2107.04034, 2021.
- Reinforcement learning for robust parameterized locomotion control of bipedal robots. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 2811–2817. IEEE, 2021.
- Robust and versatile bipedal jumping control through multi-task reinforcement learning. arXiv preprint arXiv:2302.09450, 2023.
- Physics-based character animation and human motor control. Physics of Life Reviews, 2023.
- SMPL: A skinned multi-person linear model. ACM Trans. Graphics (Proc. SIGGRAPH Asia), 34(6):248:1–248:16, October 2015.
- Universal humanoid motion representations for physics-based control. arXiv preprint arXiv:2310.04582, 2023a.
- Perpetual humanoid control for real-time simulated avatars. In International Conference on Computer Vision (ICCV), 2023b.
- Amass: Archive of motion capture as surface shapes. In The IEEE International Conference on Computer Vision (ICCV), Oct 2019. URL https://amass.is.tue.mpg.de.
- Isaac gym: High performance gpu-based physics simulation for robot learning. arXiv preprint arXiv:2108.10470, 2021.
- Learning to jump from pixels. arXiv preprint arXiv:2110.15344, 2021.
- Rapid locomotion via reinforcement learning. arXiv preprint arXiv:2205.02824, 2022.
- Dynamic walk of a biped. IJRR, 1984.
- Whole-body control of humanoid robots. Humanoid Robotics: A reference, Springer, Dordrecht, 2019.
- Deepmimic: Example-guided deep reinforcement learning of physics-based character skills. ACM Trans. Graph., 37(4):143:1–143:14, July 2018. ISSN 0730-0301. doi: 10.1145/3197517.3201311. URL http://doi.acm.org/10.1145/3197517.3201311.
- Learning agile robotic locomotion skills by imitating animals. April 2020a.
- Learning agile robotic locomotion skills by imitating animals. In Robotics: Science and Systems, 07 2020b. doi: 10.15607/RSS.2020.XVI.064.
- Amp: Adversarial motion priors for stylized physics-based character control. ACM Transactions on Graphics (ToG), 40(4):1–20, 2021.
- Ase: Large-scale reusable adversarial skill embeddings for physically simulated characters. ACM Trans. Graph., 41(4), July 2022.
- Real-world humanoid locomotion with reinforcement learning. arXiv:2303.03381, 2023.
- Learning to walk in minutes using massively parallel deep reinforcement learning. In Conference on Robot Learning, pages 91–100. PMLR, 2022.
- Curiosity-driven learning of joint locomotion and manipulation tasks. In Proceedings of The 7th Conference on Robot Learning, volume 229, pages 2594–2610. PMLR, 2023.
- Deep imitation learning for humanoid loco-manipulation through human teleoperation. In 2023 IEEE-RAS 22nd International Conference on Humanoid Robots (Humanoids), pages 1–8. IEEE, 2023.
- Blind bipedal stair traversal via sim-to-real reinforcement learning. arXiv preprint arXiv:2105.08328, 2021.
- A compliant hybrid zero dynamics controller for stable, efficient and fast bipedal walking on mabel. IJRR, 2011.
- Humanmimic: Learning natural locomotion and transitions for humanoid robot via wasserstein adversarial imitation. arXiv preprint arXiv:2309.14225, 2023.
- Calm: Conditional adversarial latent models for directable virtual characters. In ACM SIGGRAPH 2023 Conference Proceedings, SIGGRAPH ’23, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9798400701597. doi: 10.1145/3588432.3591541. URL https://doi.org/10.1145/3588432.3591541.
- Human motion diffusion model. arXiv preprint arXiv:2209.14916, 2022.
- Unicon: Universal neural controller for physics-based character motion, 2020.
- Amp in the wild: Learning robust, agile, natural legged locomotion skills. arXiv preprint arXiv:2304.10888, 2023.
- Hybrid zero dynamics of planar biped walkers. IEEE transactions on automatic control, 48(1):42–56, 2003.
- A scalable approach to control diverse behaviors for physically simulated characters. ACM Trans. Graph., 39(4), 2020. URL https://doi.org/10.1145/3386569.3392381.
- Generalized animal imitator: Agile locomotion with versatile motion prior. arXiv preprint arXiv:2310.01408, 2023a.
- Neural volumetric memory for visual locomotion control. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1430–1440, 2023b.
- Simbicon: Simple biped locomotion control. ACM Transactions on Graphics, 2007.
- Learning physically simulated tennis skills from broadcast videos. ACM Trans. Graph., 42(4), jul 2023. ISSN 0730-0301. doi: 10.1145/3592408. URL https://doi.org/10.1145/3592408.
- Robot parkour learning. In Conference on Robot Learning (CoRL), 2023.
- Xuxin Cheng (42 papers)
- Yandong Ji (8 papers)
- Junming Chen (7 papers)
- Ruihan Yang (43 papers)
- Ge Yang (49 papers)
- Xiaolong Wang (243 papers)