Accelerating Deep Neural Networks by Combining Block-Circulant Matrices and Low-Precision Weights
As a key ingredient of deep neural networks (DNNs), fully-connected (FC) layers are widely used in various artificial intelligence applications. However, there are many parameters in FC layers, so the efficient process of FC layers is restricted by memory bandwidth. In this paper, we propose a compr...
Main Authors: | Zidi Qin, Di Zhu, Xingwei Zhu, Xuan Chen, Yinghuan Shi, Yang Gao, Zhonghai Lu, Qinghong Shen, Li Li, Hongbing Pan |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2019-01-01
|
Series: | Electronics |
Subjects: | |
Online Access: | http://www.mdpi.com/2079-9292/8/1/78 |
Similar Items
-
Power-Efficient Deep Neural Network Accelerator Minimizing Global Buffer Access without Data Transfer between Neighboring Multiplier—Accumulator Units
by: Jeonghyeok Lee, et al.
Published: (2022-06-01) -
Learning Ratio Mask with Cascaded Deep Neural Networks for Echo Cancellation in Laser Monitoring Signals
by: Haitao Lang, et al.
Published: (2020-05-01) -
A Novel Automate Python Edge-to-Edge: From Automated Generation on Cloud to User Application Deployment on Edge of Deep Neural Networks for Low Power IoT Systems FPGA-Based Acceleration
by: Tarek Belabed, et al.
Published: (2021-09-01) -
A Survey of Network-Based Hardware Accelerators
by: Iouliia Skliarova
Published: (2022-03-01) -
A High-Performance and Flexible Architecture for Accelerating SDN on the MPSoC Platform
by: Meng Sha, et al.
Published: (2022-10-01)