Mitigate catastrophic forgetting in continual robot learning via data mixing
Develop methods to mix different proportions of prior-data distributions with newly collected data during fine-tuning to alleviate catastrophic forgetting in continual learning of embodied robotic policies, and establish criteria for choosing mixing ratios under diverse tasks and environments.
References
Open research problems and viable approaches include: 1) mixing different proportions of prior data distribution when fine-tuning on the latest data to alleviate catastrophic forgetting , 2) developing efficient prototypes from prior distributions or curricula for task inference in learning new tasks, 3) improving training stability and sample efficiency of online learning algorithms, 4) identifying principled ways to seamlessly incorporate large-capacity models into control frameworks, potentially through hierarchical learning or slow-fast control, for real-time inference.