Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

FairPy: A Toolkit for Evaluation of Social Biases and their Mitigation in Large Language Models (2302.05508v1)

Published 10 Feb 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Studies have shown that large pretrained LLMs exhibit biases against social groups based on race, gender etc, which they inherit from the datasets they are trained on. Various researchers have proposed mathematical tools for quantifying and identifying these biases. There have been methods proposed to mitigate such biases. In this paper, we present a comprehensive quantitative evaluation of different kinds of biases such as race, gender, ethnicity, age etc. exhibited by popular pretrained LLMs such as BERT, GPT-2 etc. and also present a toolkit that provides plug-and-play interfaces to connect mathematical tools to identify biases with large pretrained LLMs such as BERT, GPT-2 etc. and also present users with the opportunity to test custom models against these metrics. The toolkit also allows users to debias existing and custom models using the debiasing techniques proposed so far. The toolkit is available at https://github.com/HrishikeshVish/Fairpy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Hrishikesh Viswanath (13 papers)
  2. Tianyi Zhang (262 papers)
Citations (5)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub