Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient Deep Reinforcement Learning via Adaptive Policy Transfer (2002.08037v3)

Published 19 Feb 2020 in cs.LG, cs.AI, and stat.ML

Abstract: Transfer Learning (TL) has shown great potential to accelerate Reinforcement Learning (RL) by leveraging prior knowledge from past learned policies of relevant tasks. Existing transfer approaches either explicitly computes the similarity between tasks or select appropriate source policies to provide guided explorations for the target task. However, how to directly optimize the target policy by alternatively utilizing knowledge from appropriate source policies without explicitly measuring the similarity is currently missing. In this paper, we propose a novel Policy Transfer Framework (PTF) to accelerate RL by taking advantage of this idea. Our framework learns when and which source policy is the best to reuse for the target policy and when to terminate it by modeling multi-policy transfer as the option learning problem. PTF can be easily combined with existing deep RL approaches. Experimental results show it significantly accelerates the learning process and surpasses state-of-the-art policy transfer methods in terms of learning efficiency and final performance in both discrete and continuous action spaces.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Tianpei Yang (25 papers)
  2. Jianye Hao (185 papers)
  3. Zhaopeng Meng (23 papers)
  4. Zongzhang Zhang (33 papers)
  5. Yujing Hu (28 papers)
  6. Yingfeng Cheng (1 paper)
  7. Changjie Fan (79 papers)
  8. Weixun Wang (31 papers)
  9. Wulong Liu (38 papers)
  10. Zhaodong Wang (8 papers)
  11. Jiajie Peng (12 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.