Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unsupervised Learning of Effective Actions in Robotics (2404.02728v1)

Published 3 Apr 2024 in cs.RO, cs.AI, and cs.LG

Abstract: Learning actions that are relevant to decision-making and can be executed effectively is a key problem in autonomous robotics. Current state-of-the-art action representations in robotics lack proper effect-driven learning of the robot's actions. Although successful in solving manipulation tasks, deep learning methods also lack this ability, in addition to their high cost in terms of memory or training data. In this paper, we propose an unsupervised algorithm to discretize a continuous motion space and generate "action prototypes", each producing different effects in the environment. After an exploration phase, the algorithm automatically builds a representation of the effects and groups motions into action prototypes, where motions more likely to produce an effect are represented more than those that lead to negligible changes. We evaluate our method on a simulated stair-climbing reinforcement learning task, and the preliminary results show that our effect driven discretization outperforms uniformly and randomly sampled discretizations in convergence speed and maximum reward.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Affordance equivalences in robotics: A formalism. Frontiers in Neurorobotics, 12, 2018. ISSN 1662-5218. doi: 10.3389/fnbot.2018.00026. URL https://www.frontiersin.org/articles/10.3389/fnbot.2018.00026.
  2. Discovering affordances through perception and manipulation. In 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 3959–3964, 2016. doi: 10.1109/IROS.2016.7759583.
  3. Where’s the action? the pragmatic turn in cognitive science. Trends in cognitive sciences, 17(5):202–209, 2013.
  4. Manipulation learning on humanoid robots. Current Robotics Reports, 3(3):97–109, 2022.
  5. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In International conference on machine learning, pages 1861–1870. PMLR, 2018.
  6. A survey on deep reinforcement learning algorithms for robotic manipulation. Sensors, 23(7), 2023. ISSN 1424-8220. doi: 10.3390/s23073762. URL https://www.mdpi.com/1424-8220/23/7/3762.
  7. A review of robot learning for manipulation: challenges, representations, and algorithms. J. Mach. Learn. Res., 22(1), jan 2021. ISSN 1532-4435.
  8. End-to-end training of deep visuomotor policies. The Journal of Machine Learning Research, 17(1):1334–1373, 2016.
  9. James MacQueen et al. Some methods for classification and analysis of multivariate observations. In Proceedings of the fifth Berkeley symposium on mathematical statistics and probability, volume 1, pages 281–297. Oakland, CA, USA, 1967.
  10. Robust growing neural gas algorithm with application in cluster analysis. Neural networks, 17(8-9):1135–1148, 2004.
  11. Stable-baselines3: Reliable reinforcement learning implementations. Journal of Machine Learning Research, 22(268):1–8, 2021. URL http://jmlr.org/papers/v22/20-1364.html.
  12. Deep reinforcement learning for robotic hand manipulation. In 2020 International Conference on Computer, Control, Electrical, and Electronics Engineering (ICCCEEE), pages 1–5, 2021. doi: 10.1109/ICCCEEE49695.2021.9429619.
  13. To afford or not to afford: A new formalization of affordances toward affordance-based robot control. Adaptive Behavior - ADAPT BEHAV, 15:447–472, 12 2007. doi: 10.1177/1059712307084689.
  14. Bruno Siciliano and Oussama Khatib, editors. Springer Handbook of Robotics. Springer Handbooks. Springer, 2016. ISBN 978-3-319-32550-7. doi: 10.1007/978-3-319-32552-1. URL https://doi.org/10.1007/978-3-319-32552-1.
  15. Discretizing continuous action space for on-policy optimization. In Proceedings of the aaai conference on artificial intelligence, volume 34, pages 5981–5988, 2020.
  16. Gymnasium, March 2023. URL https://zenodo.org/record/8127025.
  17. Learning social affordances and using them for planning. In CogSci 2013, Berlin, Germany, July 31–August 2 2013.
  18. End-to-end reinforcement learning of robotic manipulation with robust keypoints representation, 2022.
  19. Goal Space Abstraction in Hierarchical Reinforcement Learning via Set-Based Reachability Analysis. In 2023 IEEE International Conference on Development and Learning (ICDL), pages 423–428, Macau, China, November 2023. IEEE. doi: 10.1109/ICDL55364.2023.10364473. URL https://inria.hal.science/hal-04401878.
  20. Computational models of affordance in robotics: a taxonomy and systematic classification. Adaptive Behavior, 25(5):235–271, 2017. doi: 10.1177/1059712317726357. URL https://doi.org/10.1177/1059712317726357.
  21. Action representations in robotics: A taxonomy and systematic classification. The International Journal of Robotics Research, 38(5):518–562, 2019. doi: 10.1177/0278364919835020. URL https://doi.org/10.1177/0278364919835020.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com