Latency Estimation Tool and Investigation of Neural Networks Inference on Mobile GPU

A lot of deep learning applications are desired to be run on mobile devices. Both accuracy and inference time are meaningful for a lot of them. While the number of FLOPs is usually used as a proxy for neural network latency, it may not be the best choice. In order to obtain a better approximation of...

Full description

Bibliographic Details
Main Authors: Evgeny Ponomarev, Sergey Matveev, Ivan Oseledets, Valery Glukhov
Format: Article
Language:English
Published: MDPI AG 2021-08-01
Series:Computers
Subjects:
Online Access:https://www.mdpi.com/2073-431X/10/8/104

Similar Items