AE-Qdrop: Towards Accurate and Efficient Low-Bit Post-Training Quantization for A Convolutional Neural Network
Blockwise reconstruction with adaptive rounding helps achieve acceptable 4-bit post-training quantization accuracy. However, adaptive rounding is time intensive, and the optimization space of weight elements is constrained to a binary set, thus limiting the performance of quantized models. The optim...
Main Authors: | Jixing Li, Gang Chen, Min Jin, Wenyu Mao, Huaxiang Lu |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2024-02-01
|
Series: | Electronics |
Subjects: | |
Online Access: | https://www.mdpi.com/2079-9292/13/3/644 |
Similar Items
-
Clipping-Based Post Training 8-Bit Quantization of Convolution Neural Networks for Object Detection
by: Leisheng Chen, et al.
Published: (2022-12-01) -
Training Multi-Bit Quantized and Binarized Networks with a Learnable Symmetric Quantizer
by: Phuoc Pham, et al.
Published: (2021-01-01) -
Super-Resolution Model Quantized in Multi-Precision
by: Jingyu Liu, et al.
Published: (2021-09-01) -
Two Novel Non-Uniform Quantizers with Application in Post-Training Quantization
by: Zoran Perić, et al.
Published: (2022-09-01) -
Partial Encryption of Co mpressed Image Using Threshold Quantization and AES Cipher
by: H.A.Younis, et al.
Published: (2012-01-01)