Sparse Winograd Convolutional neural networks on small-scale systolic arrays

Feng Shi, Haochen Li, Yuhe Gao, Benjamin Kuschner, Song-Chun Zhu
University of California Los Angeles, Los Angeles, USA
arXiv:1810.01973 [cs.DC], (3 Oct 2018)


   title={Sparse Winograd Convolutional neural networks on small-scale systolic arrays},

   author={Shi, Feng and Li, Haochen and Gao, Yuhe and Kuschner, Benjamin and Zhu, Song-Chun},






Download Download (PDF)   View View   Source Source   



The reconfigurability, energy-efficiency, and massive parallelism on FPGAs make them one of the best choices for implementing efficient deep learning accelerators. However, state-of-art implementations seldom consider the balance between high throughput of computation power and the ability of the memory subsystem to support it. In this paper, we implement an accelerator on FPGA by combining the sparse Winograd convolution, clusters of small-scale systolic arrays, and a tailored memory layout design. We also provide an analytical model analysis for the general Winograd convolution algorithm as a design reference. Experimental results on VGG16 show that it achieves very high computational resource utilization, 20x ~ 30x energy efficiency, and more than 5x speedup compared with the dense implementation.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2021 hgpu.org

All rights belong to the respective authors

Contact us: