29337

A Comprehensive Deep Learning Library Benchmark and Optimal Library Selection

Qiyang Zhang, Xiangying Che, Yijie Chen, Xiao Ma, Mengwei Xu, Schahram Dustdar, Xuanzhe Liu, Shangguang Wang
State Key Laboratory of Networking and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, 100876, China
IEEE Transactions on Mobile Computing, Vol. 23, No. 5, 2024

@article{zhang2023comprehensive,

   title={A comprehensive deep learning library benchmark and optimal library selection},

   author={Zhang, Qiyang and Che, Xiangying and Chen, Yijie and Ma, Xiao and Xu, Mengwei and Dustdar, Schahram and Liu, Xuanzhe and Wang, Shangguang},

   journal={IEEE Transactions on Mobile Computing},

   year={2023},

   publisher={IEEE}

}

Download Download (PDF)   View View   Source Source   Source codes Source codes

410

views

Deploying deep learning (DL) on mobile devices has been a notable trend in recent years. To support fast inference of on-device DL, DL libraries play a critical role as algorithms and hardware do. Unfortunately, no prior work ever dives deep into the ecosystem of modern DL libraries and provides quantitative results on their performance. In this paper, we first build a comprehensive benchmark that includes 6 representative DL libraries and 15 diversified DL models. Then we perform extensive experiments on 10 mobile devices, and the results reveal the current landscape of mobile DL libraries. For example, we find that the best-performing DL library is severely fragmented across different models and hardware, and the gap between DL libraries can be rather huge. In fact, the impacts of DL libraries can overwhelm the optimizations from algorithms or hardware, e.g., model quantization and GPU/DSP-based heterogeneous computing. Motivated by the fragmented performance of DL libraries across models and hardware, we propose an effective DL Library selection framework to obtain the optimal library on a new dataset that has been created. We evaluate the DL Library selection algorithm, and the results show that the framework at it can improve the prediction accuracy by about 10% than benchmark approaches on average.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2024 hgpu.org

All rights belong to the respective authors

Contact us: