Convolution Smooth: A Post-Training Quantization Method for Convolutional Neural Networks
Convolutional neural network (CNN) quantization is an efficient model compression technique primarily used for accelerating inference and optimizing resources. However, existing methods often apply different quantization strategies to activations and weights, without considering their interplay. To...
Saved in:
| Main Authors: | Yongyuan Chen, Zhendao Wang |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10955493/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
ClipQ: Clipping Optimization for the Post-Training Quantization of Convolutional Neural Network
by: Yiming Chen, et al.
Published: (2025-04-01) -
Quantized convolutional neural networks: a hardware perspective
by: Li Zhang, et al.
Published: (2025-07-01) -
Smoothed per-tensor weight quantization: a robust solution for neural network deployment
by: Xin Chang
Published: (2025-07-01) -
Quantized Convolutional Neural Networks Robustness under Perturbation [version 1; peer review: 2 approved]
by: Guy Kember, et al.
Published: (2025-04-01) -
Enhanced Vector Quantization for Embedded Machine Learning: A Post-Training Approach With Incremental Clustering
by: Thommas K. S. Flores, et al.
Published: (2025-01-01)