HACScale : hardware-aware compound scaling for resource-efficient DNNs
Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this pape...
Main Authors: | Kong, Hao, Liu, Di, Luo, Xiangzhong, Liu, Weichen, Subramaniam, Ravi |
---|---|
Other Authors: | School of Computer Science and Engineering |
Format: | Conference Paper |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/155808 |
Similar Items
-
Designing efficient DNNs via hardware-aware neural architecture search and beyond
by: Luo, Xiangzhong, et al.
Published: (2022) -
HSCoNAS : hardware-software co-design of efficient DNNs via neural architecture search
by: Luo, Xiangzhong, et al.
Published: (2022) -
Towards efficient convolutional neural network for embedded hardware via multi-dimensional pruning
by: Kong, Hao, et al.
Published: (2023) -
On hardware-aware design and optimization of edge intelligence
by: Huai, Shuo, et al.
Published: (2023) -
SurgeNAS: a comprehensive surgery on hardware-aware differentiable neural architecture search
by: Luo, Xiangzhong, et al.
Published: (2023)