ORCID
- Kelefouras, Vasilios: 0000-0001-9591-913X
Abstract
In this article, a new method is provided for accelerating the execution of convolution layers in Deep Neural Networks. This research work provides the theoretical background to efficiently design and implement the convolution layers on x86/x64 CPUs, based on the target layer parameters, quantization level and hardware architecture. The proposed work is general and can be applied to other processor families too, e.g., Arm. The proposed work achieves high speedup values over the state of the art, which is Intel oneDNN library, by applying compiler optimizations, such as vectorization, register blocking and loop tiling, in a more efficient way. This is achieved by developing an analytical modelling approach for finding the optimization parameters. A thorough experimental evaluation has been applied on two Intel CPU platforms, for DenseNet-121, ResNet-50 and SqueezeNet (including 112 different convolution layers), and for both FP32 and int8 input/output tensors (quantization). The experimental results show that the convolution layers of the aforementioned models are executed from x1.1 up to x7.2 times faster.
Publication Date
2023-10-04
Publication Title
IEEE Transactions on Parallel and Distributed Systems
ISSN
1558-2183
Embargo Period
2023-10-25
Organisational Unit
School of Engineering, Computing and Mathematics
Recommended Citation
Kelefouras, V., & Keramidas, G. (2023) 'Design and Implementation of Deep Learning 2D Convolutions on modern CPUs', IEEE Transactions on Parallel and Distributed Systems, . Retrieved from https://pearl.plymouth.ac.uk/secam-research/1310