Integrating Model-based and Model-free Reinforcement Learning for Peg-in-hole Assembly

Establish integrated reinforcement learning frameworks that combine model-based and model-free methods for peg-in-hole assembly, including accurate learning of transition dynamics and mechanisms to balance planning with the learned model against direct interaction with the real environment.

Background

The survey outlines complementary strengths of model-free RL (flexibility, ease of implementation) and model-based RL (sample efficiency, fast convergence), and reviews recent attempts to combine them.

The authors explicitly pose the open question of how to combine these approaches effectively, highlighting two key challenges: learning accurate dynamics models and balancing model-based planning with direct environment interaction during policy learning.

References

Open questions in the field of robotic peg-in-hole assembly? How can model-based and model-free RL algorithms by combined?

Compare Contact Model-based Control and Contact Model-free Learning: A Survey of Robotic Peg-in-hole Assembly Strategies (1904.05240 - Xu et al., 2019) in Section 5.1, Open questions in the field of robotic peg-in-hole assembly?, Subsubsection: How can model-based and model-free RL algorithms by combined?