Kernel Quantization for Efficient Network Compression
This paper presents a novel network compression framework, <bold>Kernel Quantization</bold> (<bold><italic>KQ</italic></bold>), targeting to efficiently convert any pre-trained full-precision convolutional neural network (CNN) model into a low-precision version wi...
Main Authors: | Zhongzhi Yu, Yemin Shi |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2022-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9672186/ |
Similar Items
-
HLQ: Hardware-Friendly Logarithmic Quantization Aware Training for Power-Efficient Low-Precision CNN Models
by: Dahun Choi, et al.
Published: (2024-01-01) -
Neuron-by-Neuron Quantization for Efficient Low-Bit QNN Training
by: Artem Sher, et al.
Published: (2023-04-01) -
CANET: Quantized Neural Network Inference With 8-bit Carry-Aware Accumulator
by: Jingxuan Yang, et al.
Published: (2024-01-01) -
Efficient Weights Quantization of Convolutional Neural Networks Using Kernel Density Estimation based Non-uniform Quantizer
by: Sanghyun Seo, et al.
Published: (2019-06-01) -
A Convolutional Neural Network-Based Quantization Method for Block Compressed Sensing of Images
by: Jiulu Gong, et al.
Published: (2024-05-01)