Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Effective Random Test Generation for Deep Learning Compilers (2302.00842v2)

Published 2 Feb 2023 in cs.SE

Abstract: Deep learning compilers help address difficulties of deploying deep learning models on diverse types of hardware. Testing deep learning compilers is highly crucial, because they are impacting countless AI applications that use them for model optimization and deployment. To test deep learning compilers, random testing, being popularly used for compiler testing practices, faces the challenge of generating semantically valid test inputs, i.e., deep learning models that satisfy the semantic model specifications (in short as semantic specifications). To tackle this challenge, in this paper, we propose a novel approach named Isra, including a domain-specific constraint solver that resolves the constraints from the semantic specifications without backtracking. We implement and apply our approach on three popular real-world deep learning compilers including TVM, Glow, and a commercial compiler. The evaluation results show that Isra is more effective than the state-of-the-art approaches and the baseline approaches on constructing valid test inputs for compiler-bug detection, and Isra successfully finds 24 previously unknown bugs in released versions of the three compilers. These results indicate effectiveness and practical value of Isra.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Luyao Ren (5 papers)
  2. Yingfei Xiong (33 papers)
  3. Li Zhang (693 papers)
  4. Guoyue Jiang (2 papers)
  5. Tao Xie (117 papers)
  6. Ziheng Wang (48 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.