Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DWM: A Decomposable Winograd Method for Convolution Acceleration (2002.00552v1)

Published 3 Feb 2020 in cs.LG, cs.CV, and eess.IV

Abstract: Winograd's minimal filtering algorithm has been widely used in Convolutional Neural Networks (CNNs) to reduce the number of multiplications for faster processing. However, it is only effective on convolutions with kernel size as 3x3 and stride as 1, because it suffers from significantly increased FLOPs and numerical accuracy problem for kernel size larger than 3x3 and fails on convolution with stride larger than 1. In this paper, we propose a novel Decomposable Winograd Method (DWM), which breaks through the limitation of original Winograd's minimal filtering algorithm to a wide and general convolutions. DWM decomposes kernels with large size or large stride to several small kernels with stride as 1 for further applying Winograd method, so that DWM can reduce the number of multiplications while keeping the numerical accuracy. It enables the fast exploring of larger kernel size and larger stride value in CNNs for high performance and accuracy and even the potential for new CNNs. Comparing against the original Winograd, the proposed DWM is able to support all kinds of convolutions with a speedup of ~2, without affecting the numerical accuracy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Di Huang (203 papers)
  2. Xishan Zhang (22 papers)
  3. Rui Zhang (1138 papers)
  4. Tian Zhi (3 papers)
  5. Deyuan He (1 paper)
  6. Jiaming Guo (37 papers)
  7. Chang Liu (865 papers)
  8. Qi Guo (237 papers)
  9. Zidong Du (41 papers)
  10. Shaoli Liu (14 papers)
  11. Tianshi Chen (45 papers)
  12. Yunji Chen (51 papers)
Citations (24)

Summary

We haven't generated a summary for this paper yet.