Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adapting Language Models for Non-Parallel Author-Stylized Rewriting (1909.09962v3)

Published 22 Sep 2019 in cs.CL and cs.LG

Abstract: Given the recent progress in LLMing using Transformer-based neural models and an active interest in generating stylized text, we present an approach to leverage the generalization capabilities of a LLM to rewrite an input text in a target author's style. Our proposed approach adapts a pre-trained LLM to generate author-stylized text by fine-tuning on the author-specific corpus using a denoising autoencoder (DAE) loss in a cascaded encoder-decoder framework. Optimizing over DAE loss allows our model to learn the nuances of an author's style without relying on parallel data, which has been a severe limitation of the previous related works in this space. To evaluate the efficacy of our approach, we propose a linguistically-motivated framework to quantify stylistic alignment of the generated text to the target author at lexical, syntactic and surface levels. The evaluation framework is both interpretable as it leads to several insights about the model, and self-contained as it does not rely on external classifiers, e.g. sentiment or formality classifiers. Qualitative and quantitative assessment indicates that the proposed approach rewrites the input text with better alignment to the target style while preserving the original content better than state-of-the-art baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Bakhtiyar Syed (1 paper)
  2. Gaurav Verma (34 papers)
  3. Balaji Vasan Srinivasan (33 papers)
  4. Anandhavelu Natarajan (9 papers)
  5. Vasudeva Varma (47 papers)
Citations (43)

Summary

We haven't generated a summary for this paper yet.