Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

N-Critics: Self-Refinement of Large Language Models with Ensemble of Critics (2310.18679v2)

Published 28 Oct 2023 in cs.CL, cs.AI, and cs.LG

Abstract: We propose a self-correction mechanism for LLMs to mitigate issues such as toxicity and fact hallucination. This method involves refining model outputs through an ensemble of critics and the model's own feedback. Drawing inspiration from human behavior, we explore whether LLMs can emulate the self-correction process observed in humans who often engage in self-reflection and seek input from others to refine their understanding of complex topics. Our approach is model-agnostic and can be applied across various domains to enhance trustworthiness by addressing fairness, bias, and robustness concerns. We consistently observe performance improvements in LLMs for reducing toxicity and correcting factual errors.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Sajad Mousavi (26 papers)
  2. Desik Rengarajan (11 papers)
  3. Vineet Gundecha (16 papers)
  4. Ashwin Ramesh Babu (20 papers)
  5. Avisek Naug (15 papers)
  6. Antonio Guillen (11 papers)
  7. Soumyendu Sarkar (17 papers)
  8. Ricardo Luna GutiƩrrez (2 papers)
Citations (4)