Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GWLZ: A Group-wise Learning-based Lossy Compression Framework for Scientific Data (2404.13470v1)

Published 20 Apr 2024 in cs.DC and cs.AI

Abstract: The rapid expansion of computational capabilities and the ever-growing scale of modern HPC systems present formidable challenges in managing exascale scientific data. Faced with such vast datasets, traditional lossless compression techniques prove insufficient in reducing data size to a manageable level while preserving all information intact. In response, researchers have turned to error-bounded lossy compression methods, which offer a balance between data size reduction and information retention. However, despite their utility, these compressors employing conventional techniques struggle with limited reconstruction quality. To address this issue, we draw inspiration from recent advancements in deep learning and propose GWLZ, a novel group-wise learning-based lossy compression framework with multiple lightweight learnable enhancer models. Leveraging a group of neural networks, GWLZ significantly enhances the decompressed data reconstruction quality with negligible impact on the compression efficiency. Experimental results on different fields from the Nyx dataset demonstrate remarkable improvements by GWLZ, achieving up to 20% quality enhancements with negligible overhead as low as 0.0003x.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Wenqi Jia (14 papers)
  2. Sian Jin (32 papers)
  3. Jinzhen Wang (2 papers)
  4. Wei Niu (68 papers)
  5. Dingwen Tao (60 papers)
  6. Miao Yin (25 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.