A Reinforcement Learning Based Controller to Minimize Forces on the Crutches of a Lower-Limb Exoskeleton (2402.00135v1)
Abstract: Metabolic energy consumption of a powered lower-limb exoskeleton user mainly comes from the upper body effort since the lower body is considered to be passive. However, the upper body effort of the users is largely ignored in the literature when designing motion controllers. In this work, we use deep reinforcement learning to develop a locomotion controller that minimizes ground reaction forces (GRF) on crutches. The rationale for minimizing GRF is to reduce the upper body effort of the user. Accordingly, we design a model and a learning framework for a human-exoskeleton system with crutches. We formulate a reward function to encourage the forward displacement of a human-exoskeleton system while satisfying the predetermined constraints of a physical robot. We evaluate our new framework using Proximal Policy Optimization, a state-of-the-art deep reinforcement learning (RL) method, on the MuJoCo physics simulator with different hyperparameters and network architectures over multiple trials. We empirically show that our learning model can generate joint torques based on the joint angle, velocities, and the GRF on the feet and crutch tips. The resulting exoskeleton model can directly generate joint torques from states in line with the RL framework. Finally, we empirically show that policy trained using our method can generate a gait with a 35% reduction in GRF with respect to the baseline.
- Ouyang, W., Y. Wang, S. Han, Z. Jin and P. Weng, “Improving Generalization of Deep Reinforcement Learning-based TSP Solvers”, 2021 IEEE Symposium Series on Computational Intelligence (SSCI), Orlando, USA, pp. 01-08, 2021.
- Liu, M., Wang, D., Huang, H., “Development of an environment-aware locomotion mode recognition system for powered lower limb prostheses”, IEEE Transactions on Neural Systems and Rehabilitation Engineering, 24(4), pp. 434-443., 2015.
- Wen, Y., Si, J., Brandt, A., Gao, X., & Huang, H. H., “Online reinforcement learning control for the personalization of a robotic knee prosthesis”, IEEE transactions on cybernetics, 50(6), pp. 2346-2356., 2019
- Lillicrap, T.P., J.J. Hunt, A. Pritzel, N.M. Heess, T. Erez, Y. Tassa, D. Silver and D. Wierstra, “Continuous control with deep reinforcement learning”, arXiv:1509.02971, 2015.
- Mnih, V., K. Kavukcuoglu, D. Silver, A.A. Rusu, J. Veness, M.G. Bellemare, A. Graves, M. Riedmiller, A.K. Fidjeland, G. Ostrovski and S. Petersen, “Human-level control through deep reinforcement learning”, Nature, Vol. 518, No. 7540, pp. 529-533, 2015.
- M.C. Yildirim, A.T. Kansizoglu, S. Emre, M. Derman, S. Coruk, A.F. Soliman, P. Sendur, and B. Ugurlu, “Co-Ex: A Torque-Controllable Lower Body Exoskeleton for Dependable Human-Robot Co-Existence”, in Proc. of the IEEE International Conference on Rehabilitation Robotics (ICORR), Toronto, Canada, pp. 605-610, 2019
- Fujimoto, S., H. Hoof and D. Meger, “Addressing function approximation error in actor-critic methods”, International Conference on Machine Learning, Stockholm, Sweden, Vol.80, pp. 1582-1591, 2018.
- Tirupachuri Y., L. Rapetti, C. Latella, R. Grieco, D. Ferigo, K. Darvish, “human-gazebo”, GitHub repository, 2019, https://github.com/robotology/human-gazebo, accessed on December 11, 2022.
- Latella C. and L. Rapetti, “human-model-generator”, GitHub repository, 2020,https://github.com/ami-iit/human-model-generator, accessed on December 11, 2022.
- Schulman, J., F. Wolski, P. Dhariwal, A. Radford and O. Klimov, “Proximal policy optimization algorithms”, arXiv:1707.06347, 2017.
- Chou P.W., D. Maturana, and S. Scherer, “Improving stochastic policy gradients in continuous control with deep reinforcement learning using the beta distribution”, International Conference on Machine Learning, Sydney, Australia, pp. 834–843, 2017.
- Dhariwal, P., C. Hesse, O. Klimov, A. Nichol, M. Plappert, A. Radford, J. Schulman, S. Sidor, Y. Wu and P. Zhokhov, “OpenAI Baselines”, GitHub repository, 2017,https://github.com/openai/baselines, accessed on December 11, 2022.