Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MLModelScope: A Distributed Platform for Model Evaluation and Benchmarking at Scale (2002.08295v1)

Published 19 Feb 2020 in cs.DC, cs.LG, and stat.ML

Abstract: Machine Learning (ML) and Deep Learning (DL) innovations are being introduced at such a rapid pace that researchers are hard-pressed to analyze and study them. The complicated procedures for evaluating innovations, along with the lack of standard and efficient ways of specifying and provisioning ML/DL evaluation, is a major "pain point" for the community. This paper proposes MLModelScope, an open-source, framework/hardware agnostic, extensible and customizable design that enables repeatable, fair, and scalable model evaluation and benchmarking. We implement the distributed design with support for all major frameworks and hardware, and equip it with web, command-line, and library interfaces. To demonstrate MLModelScope's capabilities we perform parallel evaluation and show how subtle changes to model evaluation pipeline affects the accuracy and HW/SW stack choices affect performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Abdul Dakkak (11 papers)
  2. Cheng Li (1094 papers)
  3. Jinjun Xiong (118 papers)
  4. Wen-mei Hwu (62 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.