Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Direct Judgement Preference Optimization (2409.14664v2)

Published 23 Sep 2024 in cs.CL
Direct Judgement Preference Optimization

Abstract: Auto-evaluation is crucial for assessing response quality and offering feedback for model development. Recent studies have explored training LLMs as generative judges to evaluate and critique other models' outputs. In this work, we investigate the idea of learning from both positive and negative data with preference optimization to enhance the evaluation capabilities of LLM judges across an array of different use cases. We achieve this by employing three approaches to collect the preference pairs for different use cases, each aimed at improving our generative judge from a different perspective. Our comprehensive study over a wide range of benchmarks demonstrates the effectiveness of our method. In particular, our generative judge achieves the best performance on 10 out of 13 benchmarks, outperforming strong baselines like GPT-4o and specialized judge models. Further analysis show that our judge model robustly counters inherent biases such as position and length bias, flexibly adapts to any evaluation protocol specified by practitioners, and provides helpful language feedback for improving downstream generator models.

Direct Judgment Preference Optimization

The paper "Direct Judgement Preference Optimization" by Wang et al. presents a sophisticated approach to enhancing the evaluation capabilities of LLM judges. By leveraging a method that trains these models through preference optimization using both positive and negative examples, the paper addresses several critical challenges associated with model evaluation. This essay will provide a detailed overview of the paper, focusing on the methodologies employed, the experimental results obtained, and the implications for future developments in AI.

The primary focus of the paper is on auto-evaluation, which is pivotal for assessing the quality of model responses and providing feedback for further model development. Human evaluations, while accurate, are expensive and unscalable. Therefore, the utilization of LLMs as generative judges—models that not only assess but also critique other models' outputs—is presented as a feasible alternative.

Methodology

The approach proposed by Wang et al. builds on the existing supervised fine-tuning (SFT) techniques by integrating direct judgement preference optimization (DPO). The essence of this methodology is to train generative judges not just on correctly labeled data but also on data where incorrect judgments are explicitly noted and avoided.

Data Collection and Tasks

Three primary training tasks are identified:

  1. Chain-of-Thought (CoT) Critique: Here, an LLM is prompted to provide detailed evaluations and final judgments on responses. Positive examples are those where the judgment aligns with ground-truth annotations, while negative examples do not. This structure aims to improve the model's reasoning capabilities.
  2. Standard Judgement: This task involves generating judgements without the detailed CoT critiques, focusing on directly learning from correct and incorrect judgments to refine the model's decision-making precision.
  3. Response Deduction: An auxiliary task where the model is trained to deduce the original responses given the model's evaluations. This reverse-engineering process bolsters the model’s understanding of what constitutes good or bad responses.

The model is trained using these preference pairs across a variety of datasets, ensuring a wide coverage of evaluation tasks like single rating, pairwise comparison, and classification. This comprehensive dataset aids in the generalizability of the trained judge model across different domains.

Experimental Results

Experiments demonstrate the efficacy of the proposed method across 13 benchmarks encompassing safety, reasoning, and instructional tasks. The results are striking:

  • The largest model achieves best performance in 10 out of 13 benchmarks, outperforming established models like GPT-4o and other specialized judge models.
  • The models show robustness against biases such as position and length bias.
  • The judge models are adaptable to various evaluation protocols and provide actionable feedback for improving downstream models.

Bias Analysis and Prompt Robustness

The paper also explores the inherent biases present in evaluation models. The trained judge models, referred to as SFR-Judges, show notable reductions in biases compared with other existing models. The analysis includes a breakdown of performance across different bias categories in EvalBiasBench and pairwise comparison consistency metrics.

Additionally, the flexibility of SFR-Judges in adapting to different prompting styles—whether task-specific or general—ensures their broad applicability. The models maintained their performance levels across varying prompts, showcasing their robustness and reliability.

Implications and Future Directions

The implications of this research are far-reaching:

  • Practical Impact: The ability to provide reliable, bias-mitigated evaluations can significantly streamline the model development lifecycle, making the training and refinement of LLMs more efficient and cost-effective.
  • Theoretical Advancements: The integration of DPO in training generative judges represents a notable shift from purely SFT methodologies, highlighting the benefits of learning from both correct and incorrect examples.
  • Future Developments: Potential future work might explore scaling these models further, integrating more diverse datasets, and refining the DPO techniques to handle increasingly complex evaluation tasks.

In conclusion, the paper by Wang et al. offers a comprehensive and effective strategy for enhancing the capabilities of LLM judges. By introducing a multi-faceted training approach that includes both reasoning and judgement-specific tasks, the proposed models achieve remarkable performance and adaptability. This work not only advances the state of auto-evaluation in AI but also sets a new benchmark for training and utilizing generative judge models in future AI development endeavors.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Peifeng Wang (14 papers)
  2. Austin Xu (12 papers)
  3. Yilun Zhou (28 papers)
  4. Caiming Xiong (337 papers)
  5. Shafiq Joty (187 papers)
Citations (6)