Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-Sample Training for Neural Image Compression (2209.13834v1)

Published 28 Sep 2022 in cs.CV

Abstract: This paper considers the problem of lossy neural image compression (NIC). Current state-of-the-art (sota) methods adopt uniform posterior to approximate quantization noise, and single-sample pathwise estimator to approximate the gradient of evidence lower bound (ELBO). In this paper, we propose to train NIC with multiple-sample importance weighted autoencoder (IWAE) target, which is tighter than ELBO and converges to log likelihood as sample size increases. First, we identify that the uniform posterior of NIC has special properties, which affect the variance and bias of pathwise and score function estimators of the IWAE target. Moreover, we provide insights on a commonly adopted trick in NIC from gradient variance perspective. Based on those analysis, we further propose multiple-sample NIC (MS-NIC), an enhanced IWAE target for NIC. Experimental results demonstrate that it improves sota NIC methods. Our MS-NIC is plug-and-play, and can be easily extended to other neural compression tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Tongda Xu (31 papers)
  2. Yan Wang (734 papers)
  3. Dailan He (25 papers)
  4. Chenjian Gao (7 papers)
  5. Han Gao (78 papers)
  6. Kunzan Liu (8 papers)
  7. Hongwei Qin (38 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.