Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Interpretable Unified Language Checking (2304.03728v1)

Published 7 Apr 2023 in cs.CL

Abstract: Despite recent concerns about undesirable behaviors generated by LLMs, including non-factual, biased, and hateful language, we find LLMs are inherent multi-task language checkers based on their latent representations of natural and social knowledge. We present an interpretable, unified, language checking (UniLC) method for both human and machine-generated language that aims to check if language input is factual and fair. While fairness and fact-checking tasks have been handled separately with dedicated models, we find that LLMs can achieve high performance on a combination of fact-checking, stereotype detection, and hate speech detection tasks with a simple, few-shot, unified set of prompts. With the ``1/2-shot'' multi-task language checking method proposed in this work, the GPT3.5-turbo model outperforms fully supervised baselines on several language tasks. The simple approach and results suggest that based on strong latent knowledge representations, an LLM can be an adaptive and explainable tool for detecting misinformation, stereotypes, and hate speech.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Tianhua Zhang (10 papers)
  2. Hongyin Luo (31 papers)
  3. Yung-Sung Chuang (37 papers)
  4. Wei Fang (98 papers)
  5. Luc Gaitskell (1 paper)
  6. Thomas Hartvigsen (46 papers)
  7. Xixin Wu (85 papers)
  8. Danny Fox (2 papers)
  9. Helen Meng (204 papers)
  10. James Glass (173 papers)
Citations (17)
Github Logo Streamline Icon: https://streamlinehq.com