<i>Nebula</i>: A Scalable and Flexible Accelerator for DNN Multi-Branch Blocks on Embedded Systems
Deep neural networks (DNNs) are widely used in many artificial intelligence applications; many specialized DNN-inference accelerators have been proposed. However, existing DNN accelerators rely heavily on certain types of DNN operations (such as Conv, FC, and ReLU, etc.), which are either less used...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-02-01
|
Series: | Electronics |
Subjects: | |
Online Access: | https://www.mdpi.com/2079-9292/11/4/505 |
Summary: | Deep neural networks (DNNs) are widely used in many artificial intelligence applications; many specialized DNN-inference accelerators have been proposed. However, existing DNN accelerators rely heavily on certain types of DNN operations (such as Conv, FC, and ReLU, etc.), which are either less used or likely to become out of date in future, posing challenges of flexibility and compatibility to existing work. This paper designs a flexible DNN accelerator from a more generic perspective rather than speeding up certain types of DNN operations. Our proposed <i>Nebula</i> exploits the width property of DNNs and gains a significant improvement in system throughput and energy efficiency over multi-branch architectures. <i>Nebula</i> is a first-of-its-kind framework for multi-branch DNNs. |
---|---|
ISSN: | 2079-9292 |