HACScale : hardware-aware compound scaling for resource-efficient DNNs
Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this pape...
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Conference Paper |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/155808 |
_version_ | 1811688630653026304 |
---|---|
author | Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi |
author2 | School of Computer Science and Engineering |
author_facet | School of Computer Science and Engineering Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi |
author_sort | Kong, Hao |
collection | NTU |
description | Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this paper, we propose HACScale, a hardware-aware model scaling strategy to fully exploit hardware resources for higher accuracy. In HACScale, different dimensions of DNNs are jointly scaled with consideration of their contributions to hardware utilization and accuracy. To improve the efficiency of width scaling, we introduce importance-aware width scaling in HACScale, which computes the importance of each layer to the accuracy and scales each layer accordingly to optimize the trade-off between accuracy and model parameters. Experiments show that HACScale improves the hardware utilization by 1.92× on ImageNet, as a result, it achieves 2.41% accuracy improvement with a negligible latency increase of 0.6%. On CIFAR-10, HACScale improves the accuracy by 2.23% with only 6.5% latency growth. |
first_indexed | 2024-10-01T05:35:16Z |
format | Conference Paper |
id | ntu-10356/155808 |
institution | Nanyang Technological University |
language | English |
last_indexed | 2024-10-01T05:35:16Z |
publishDate | 2022 |
record_format | dspace |
spelling | ntu-10356/1558082023-12-15T03:06:01Z HACScale : hardware-aware compound scaling for resource-efficient DNNs Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi School of Computer Science and Engineering 2022 27th Asia and South Pacific Design Automation Conference (ASP-DAC) HP-NTU Digital Manufacturing Corporate Lab Engineering::Computer science and engineering Deep Learning Design Automation Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this paper, we propose HACScale, a hardware-aware model scaling strategy to fully exploit hardware resources for higher accuracy. In HACScale, different dimensions of DNNs are jointly scaled with consideration of their contributions to hardware utilization and accuracy. To improve the efficiency of width scaling, we introduce importance-aware width scaling in HACScale, which computes the importance of each layer to the accuracy and scales each layer accordingly to optimize the trade-off between accuracy and model parameters. Experiments show that HACScale improves the hardware utilization by 1.92× on ImageNet, as a result, it achieves 2.41% accuracy improvement with a negligible latency increase of 0.6%. On CIFAR-10, HACScale improves the accuracy by 2.23% with only 6.5% latency growth. Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This study is supported under the RIE2020 Industry Alignment Fund – Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner, HP Inc., through the HP-NTU Digital Manufacturing Corporate Lab. This work is also partially supported by NTU NAP M4082282 and SUG M4082087, Singapore. 2022-03-22T02:32:35Z 2022-03-22T02:32:35Z 2022 Conference Paper Kong, H., Liu, D., Luo, X., Liu, W. & Subramaniam, R. (2022). HACScale : hardware-aware compound scaling for resource-efficient DNNs. 2022 27th Asia and South Pacific Design Automation Conference (ASP-DAC), 708-713. https://dx.doi.org/10.1109/ASP-DAC52403.2022.9712593 9781665421355 https://hdl.handle.net/10356/155808 10.1109/ASP-DAC52403.2022.9712593 2-s2.0-85122940747 708 713 en M4082282 M4082087 10.21979/N9/KSXK4T © 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: https://doi.org/10.1109/ASP-DAC52403.2022.9712593. application/pdf |
spellingShingle | Engineering::Computer science and engineering Deep Learning Design Automation Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title | HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_full | HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_fullStr | HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_full_unstemmed | HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_short | HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_sort | hacscale hardware aware compound scaling for resource efficient dnns |
topic | Engineering::Computer science and engineering Deep Learning Design Automation |
url | https://hdl.handle.net/10356/155808 |
work_keys_str_mv | AT konghao hacscalehardwareawarecompoundscalingforresourceefficientdnns AT liudi hacscalehardwareawarecompoundscalingforresourceefficientdnns AT luoxiangzhong hacscalehardwareawarecompoundscalingforresourceefficientdnns AT liuweichen hacscalehardwareawarecompoundscalingforresourceefficientdnns AT subramaniamravi hacscalehardwareawarecompoundscalingforresourceefficientdnns |