AI-Driven Performance Modeling for AI Inference Workloads
Deep Learning (DL) is moving towards deploying workloads not only in cloud datacenters, but also to the local devices. Although these are mostly limited to inference tasks, it still widens the range of possible target architectures significantly. Additionally, these new targets usually come with dra...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-07-01
|
Series: | Electronics |
Subjects: | |
Online Access: | https://www.mdpi.com/2079-9292/11/15/2316 |
_version_ | 1797414232234393600 |
---|---|
author | Max Sponner Bernd Waschneck Akash Kumar |
author_facet | Max Sponner Bernd Waschneck Akash Kumar |
author_sort | Max Sponner |
collection | DOAJ |
description | Deep Learning (DL) is moving towards deploying workloads not only in cloud datacenters, but also to the local devices. Although these are mostly limited to inference tasks, it still widens the range of possible target architectures significantly. Additionally, these new targets usually come with drastically reduced computation performance and memory sizes compared to the traditionally used architectures—and put the key optimization focus on the efficiency as they often depend on batteries. To help developers quickly estimate the performance of a neural network during its design phase, performance models could be used. However, these models are expensive to implement as they require in-depth knowledge about the hardware architecture and the used algorithms. Although AI-based solutions exist, these either require large datasets that are difficult to collect on the low-performance targets and/or limited to a small number of target platforms and metrics. Our solution exploits the block-based structure of neural networks, as well as the high similarity in the typically used layer configurations across neural networks, enabling the training of accurate models on significantly smaller datasets. In addition, our solution is not limited to a specific architecture or metric. We showcase the feasibility of the solution on a set of seven devices from four different hardware architectures, and with up to three performance metrics per target—including the power consumption and memory footprint. Our tests have shown that the solution achieved an error of less than 1 ms (2.6%) in latency, 0.12 J (4%) in energy consumption and 11 MiB (1.5%) in memory allocation for the whole network inference prediction, while being up to five orders of magnitude faster than a benchmark. |
first_indexed | 2024-03-09T05:30:06Z |
format | Article |
id | doaj.art-402f8a9fe45f40f99832e9adfb9a8db5 |
institution | Directory Open Access Journal |
issn | 2079-9292 |
language | English |
last_indexed | 2024-03-09T05:30:06Z |
publishDate | 2022-07-01 |
publisher | MDPI AG |
record_format | Article |
series | Electronics |
spelling | doaj.art-402f8a9fe45f40f99832e9adfb9a8db52023-12-03T12:33:11ZengMDPI AGElectronics2079-92922022-07-011115231610.3390/electronics11152316AI-Driven Performance Modeling for AI Inference WorkloadsMax Sponner0Bernd Waschneck1Akash Kumar2Infineon Technologies Dresden GmbH & Co. KG, 01099 Dresden, GermanyInfineon Technologies Dresden GmbH & Co. KG, 01099 Dresden, GermanyCenter for Advancing Electronics Dresden (CFAED), Technical University (TU) Dresden, 01062 Dresden, GermanyDeep Learning (DL) is moving towards deploying workloads not only in cloud datacenters, but also to the local devices. Although these are mostly limited to inference tasks, it still widens the range of possible target architectures significantly. Additionally, these new targets usually come with drastically reduced computation performance and memory sizes compared to the traditionally used architectures—and put the key optimization focus on the efficiency as they often depend on batteries. To help developers quickly estimate the performance of a neural network during its design phase, performance models could be used. However, these models are expensive to implement as they require in-depth knowledge about the hardware architecture and the used algorithms. Although AI-based solutions exist, these either require large datasets that are difficult to collect on the low-performance targets and/or limited to a small number of target platforms and metrics. Our solution exploits the block-based structure of neural networks, as well as the high similarity in the typically used layer configurations across neural networks, enabling the training of accurate models on significantly smaller datasets. In addition, our solution is not limited to a specific architecture or metric. We showcase the feasibility of the solution on a set of seven devices from four different hardware architectures, and with up to three performance metrics per target—including the power consumption and memory footprint. Our tests have shown that the solution achieved an error of less than 1 ms (2.6%) in latency, 0.12 J (4%) in energy consumption and 11 MiB (1.5%) in memory allocation for the whole network inference prediction, while being up to five orders of magnitude faster than a benchmark.https://www.mdpi.com/2079-9292/11/15/2316performance modelingmachine learningregression models |
spellingShingle | Max Sponner Bernd Waschneck Akash Kumar AI-Driven Performance Modeling for AI Inference Workloads Electronics performance modeling machine learning regression models |
title | AI-Driven Performance Modeling for AI Inference Workloads |
title_full | AI-Driven Performance Modeling for AI Inference Workloads |
title_fullStr | AI-Driven Performance Modeling for AI Inference Workloads |
title_full_unstemmed | AI-Driven Performance Modeling for AI Inference Workloads |
title_short | AI-Driven Performance Modeling for AI Inference Workloads |
title_sort | ai driven performance modeling for ai inference workloads |
topic | performance modeling machine learning regression models |
url | https://www.mdpi.com/2079-9292/11/15/2316 |
work_keys_str_mv | AT maxsponner aidrivenperformancemodelingforaiinferenceworkloads AT berndwaschneck aidrivenperformancemodelingforaiinferenceworkloads AT akashkumar aidrivenperformancemodelingforaiinferenceworkloads |