Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

XFBoost: Improving Text Generation with Controllable Decoders (2202.08124v1)

Published 16 Feb 2022 in cs.CL

Abstract: Multimodal conditionality in transformer-based natural LLMs has demonstrated state-of-the-art performance in the task of product description generation. Recent approaches condition a LLM on one or more images and other textual metadata to achieve near-human performance for describing products from e-commerce stores. However, generated descriptions may exhibit degrees of inaccuracy or even contradictory claims relative to the inputs of a given product. In this paper, we propose a controllable language generation framework called Extract-Finetune-Boost (XFBoost), which addresses the problem of inaccurate low-quality inference. By using visual semantic attributes as constraints at the decoding stage of the generation process and finetuning the LLM with policy gradient techniques, the XFBoost framework is found to produce significantly more descriptive text with higher image relevancy, outperforming baselines and lowering the frequency of factually inaccurate descriptions. We further demonstrate the application of XFBoost to online learning wherein human-in-the-loop critics improve LLMs with active feedback.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Xiangyu Peng (33 papers)
  2. Michael Sollami (5 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.