<i>Nebula</i>: A Scalable and Flexible Accelerator for DNN Multi-Branch Blocks on Embedded Systems

Deep neural networks (DNNs) are widely used in many artificial intelligence applications; many specialized DNN-inference accelerators have been proposed. However, existing DNN accelerators rely heavily on certain types of DNN operations (such as Conv, FC, and ReLU, etc.), which are either less used...

Full description

Bibliographic Details
Main Authors: Dawei Yang, Xinlei Li, Lizhe Qi, Wenqiang Zhang, Zhe Jiang
Format: Article
Language:English
Published: MDPI AG 2022-02-01
Series:Electronics
Subjects:
Online Access:https://www.mdpi.com/2079-9292/11/4/505
Description
Summary:Deep neural networks (DNNs) are widely used in many artificial intelligence applications; many specialized DNN-inference accelerators have been proposed. However, existing DNN accelerators rely heavily on certain types of DNN operations (such as Conv, FC, and ReLU, etc.), which are either less used or likely to become out of date in future, posing challenges of flexibility and compatibility to existing work. This paper designs a flexible DNN accelerator from a more generic perspective rather than speeding up certain types of DNN operations. Our proposed <i>Nebula</i> exploits the width property of DNNs and gains a significant improvement in system throughput and energy efficiency over multi-branch architectures. <i>Nebula</i> is a first-of-its-kind framework for multi-branch DNNs.
ISSN:2079-9292