Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Prototypical Reward Network for Data-Efficient RLHF (2406.06606v2)

Published 6 Jun 2024 in cs.CL and cs.AI

Abstract: The reward model for Reinforcement Learning from Human Feedback (RLHF) has proven effective in fine-tuning LLMs. Notably, collecting human feedback for RLHF can be resource-intensive and lead to scalability issues for LLMs and complex tasks. Our proposed framework Proto-RM leverages prototypical networks to enhance reward models under limited human feedback. By enabling stable and reliable structural learning from fewer samples, Proto-RM significantly enhances LLMs' adaptability and accuracy in interpreting human preferences. Extensive experiments on various datasets demonstrate that Proto-RM significantly improves the performance of reward models and LLMs in human feedback tasks, achieving comparable and usually better results than traditional methods, while requiring significantly less data. in data-limited scenarios. This research offers a promising direction for enhancing the efficiency of reward models and optimizing the fine-tuning of LLMs under restricted feedback conditions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jinghan Zhang (18 papers)
  2. Xiting Wang (42 papers)
  3. Yiqiao Jin (27 papers)
  4. Changyu Chen (19 papers)
  5. Xinhao Zhang (13 papers)
  6. Kunpeng Liu (54 papers)
Citations (11)
X Twitter Logo Streamline Icon: https://streamlinehq.com