Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Exploring Cross-lingual Textual Style Transfer with Large Multilingual Language Models (2206.02252v1)

Published 5 Jun 2022 in cs.CL

Abstract: Detoxification is a task of generating text in polite style while preserving meaning and fluency of the original toxic text. Existing detoxification methods are designed to work in one exact language. This work investigates multilingual and cross-lingual detoxification and the behavior of large multilingual models like in this setting. Unlike previous works we aim to make LLMs able to perform detoxification without direct fine-tuning in given language. Experiments show that multilingual models are capable of performing multilingual style transfer. However, models are not able to perform cross-lingual detoxification and direct fine-tuning on exact language is inevitable.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Daniil Moskovskiy (9 papers)
  2. Daryna Dementieva (20 papers)
  3. Alexander Panchenko (92 papers)
Citations (3)