18417

Anatomy Of High-Performance Deep Learning Convolutions On SIMD Architectures

Evangelos Georganas, Sasikanth Avancha, Kunal Banerjee, Dhiraj Kalamkar, Greg Henry, Hans Pabst, Alexander Heinecke
Intel Corporation
arXiv:1808.05567 [cs.DC], (16 Aug 2018)

@article{georganas2018anatomy,

   title={Anatomy Of High-Performance Deep Learning Convolutions On SIMD Architectures},

   author={Georganas, Evangelos and Avancha, Sasikanth and Banerjee, Kunal and Kalamkar, Dhiraj and Henry, Greg and Pabst, Hans and Heinecke, Alexander},

   year={2018},

   month={aug},

   archivePrefix={"arXiv"},

   primaryClass={cs.DC}

}

Convolution layers are prevalent in many classes of deep neural networks, including Convolutional Neural Networks (CNNs) which provide state-of-the-art results for tasks like image recognition, neural machine translation and speech recognition. The computationally expensive nature of a convolution operation has led to the proliferation of implementations including matrix-matrix multiplication formulation, and direct convolution primarily targeting GPUs. In this paper, we introduce direct convolution kernels for x86 architectures, in particular for Xeon and XeonPhi systems, which are implemented via a dynamic compilation approach. Our JIT-based implementation shows close to theoretical peak performance, depending on the setting and the CPU architecture at hand. We additionally demonstrate how these JIT-optimized kernels can be integrated into a lightweight multi-node graph execution model. This illustrates that single- and multi-node runs yield high efficiencies and high image-throughputs when executing state-of-the-art image recognition tasks on CPUs.
Rating: 3.5/5. From 2 votes.
Please wait...

* * *

* * *

HGPU group © 2010-2024 hgpu.org

All rights belong to the respective authors

Contact us: