Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Automatic Construction of a Korean Toxic Instruction Dataset for Ethical Tuning of Large Language Models (2311.18215v1)

Published 30 Nov 2023 in cs.CL

Abstract: Caution: this paper may include material that could be offensive or distressing. The advent of LLMs necessitates the development of training approaches that mitigate the generation of unethical language and aptly manage toxic user queries. Given the challenges related to human labor and the scarcity of data, we present KoTox, comprising 39K unethical instruction-output pairs. This collection of automatically generated toxic instructions refines the training of LLMs and establishes a foundational framework for improving LLMs' ethical awareness and response to various toxic inputs, promoting more secure and responsible interactions in NLP applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Sungjoo Byun (9 papers)
  2. Dongjun Jang (8 papers)
  3. Hyemi Jo (4 papers)
  4. Hyopil Shin (12 papers)
Citations (1)