Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Transformer-based Image Compression with Variable Image Quality Objectives (2309.12717v1)

Published 22 Sep 2023 in cs.CV and cs.MM

Abstract: This paper presents a Transformer-based image compression system that allows for a variable image quality objective according to the user's preference. Optimizing a learned codec for different quality objectives leads to reconstructed images with varying visual characteristics. Our method provides the user with the flexibility to choose a trade-off between two image quality objectives using a single, shared model. Motivated by the success of prompt-tuning techniques, we introduce prompt tokens to condition our Transformer-based autoencoder. These prompt tokens are generated adaptively based on the user's preference and input image through learning a prompt generation network. Extensive experiments on commonly used quality metrics demonstrate the effectiveness of our method in adapting the encoding and/or decoding processes to a variable quality objective. While offering the additional flexibility, our proposed method performs comparably to the single-objective methods in terms of rate-distortion performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Chia-Hao Kao (7 papers)
  2. Yi-Hsin Chen (37 papers)
  3. Cheng Chien (4 papers)
  4. Wei-Chen Chiu (54 papers)
  5. Wen-Hsiao Peng (39 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.