Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Conditional Entropy Coding for Efficient Video Compression (2008.09180v1)

Published 20 Aug 2020 in eess.IV, cs.CV, cs.IT, and math.IT

Abstract: We propose a very simple and efficient video compression framework that only focuses on modeling the conditional entropy between frames. Unlike prior learning-based approaches, we reduce complexity by not performing any form of explicit transformations between frames and assume each frame is encoded with an independent state-of-the-art deep image compressor. We first show that a simple architecture modeling the entropy between the image latent codes is as competitive as other neural video compression works and video codecs while being much faster and easier to implement. We then propose a novel internal learning extension on top of this architecture that brings an additional 10% bitrate savings without trading off decoding speed. Importantly, we show that our approach outperforms H.265 and other deep learning baselines in MS-SSIM on higher bitrate UVG video, and against all video codecs on lower framerates, while being thousands of times faster in decoding than deep models utilizing an autoregressive entropy model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jerry Liu (10 papers)
  2. Shenlong Wang (70 papers)
  3. Wei-Chiu Ma (46 papers)
  4. Meet Shah (8 papers)
  5. Rui Hu (96 papers)
  6. Pranaab Dhawan (2 papers)
  7. Raquel Urtasun (161 papers)
Citations (56)

Summary

We haven't generated a summary for this paper yet.