MobileNets Can Be Lossily Compressed: Neural Network Compression for Embedded Accelerators

Although neural network quantization is an imperative technology for the computation and memory efficiency of embedded neural network accelerators, simple post-training quantization incurs unacceptable levels of accuracy degradation on some important models targeting embedded systems, such as Mobile...

Full description

Bibliographic Details
Main Authors: Se-Min Lim, Sang-Woo Jun
Format: Article
Language:English
Published: MDPI AG 2022-03-01
Series:Electronics
Subjects:
Online Access:https://www.mdpi.com/2079-9292/11/6/858