Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

HPC AI500: Representative, Repeatable and Simple HPC AI Benchmarking (2102.12848v1)

Published 25 Feb 2021 in cs.PF

Abstract: Recent years witness a trend of applying large-scale distributed deep learning algorithms (HPC AI) in both business and scientific computing areas, whose goal is to speed up the training time to achieve a state-of-the-art quality. The HPC AI benchmarks accelerate the process. Unfortunately, benchmarking HPC AI systems at scale raises serious challenges. This paper presents a representative, repeatable and simple HPC AI benchmarking methodology. Among the seventeen AI workloads of AIBench Training -- by far the most comprehensive AI Training benchmarks suite -- we choose two representative and repeatable AI workloads. The selected HPC AI benchmarks include both business and scientific computing: Image Classification and Extreme Weather Analytics. To rank HPC AI systems, we present a new metric named Valid FLOPS, emphasizing both throughput performance and a target quality. The specification, source code, datasets, and HPC AI500 ranking numbers are publicly available from \url{https://www.benchcouncil.org/HPCAI500/}.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Zihan Jiang (19 papers)
  2. Wanling Gao (47 papers)
  3. Fei Tang (29 papers)
  4. Xingwang Xiong (5 papers)
  5. Lei Wang (975 papers)
  6. Chuanxin Lan (5 papers)
  7. Chunjie Luo (39 papers)
  8. Hongxiao Li (5 papers)
  9. Jianfeng Zhan (92 papers)
Citations (3)