WAPS-Quant: Low-Bit Post-Training Quantization Using Weight-Activation Product Scaling
Post-Training Quantization (PTQ) has been effectively compressing neural networks into very few bits using a limited calibration dataset. Various quantization methods utilizing second-order error have been proposed and demonstrated good performance. However, at extremely low bits, the increase in qu...
Saved in:
| Main Authors: | Geunjae Choi, Kamin Lee, Nojun Kwak |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10982219/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
by: Aozhong Zhang, et al.
Published: (2025-01-01) -
Evaluation of Information Indicators of Quantization Noise Generated by Speech Signal Analogue-Digital Conversion
by: V. K. Zheleznyak, et al.
Published: (2023-04-01) -
Synthesis of a measuring composite signal for assessing the security of speech signals during discrete-quantized transformation
by: V. K. Zheleznjak, et al.
Published: (2020-10-01) -
Burau representation of $B_4$ and quantization of the rational projective plane
by: Jouteur, Perrine
Published: (2025-03-01) -
Conditional Optimal Sets and the Quantization Coefficients for Some Uniform Distributions
by: Evans Nyanney, et al.
Published: (2025-07-01)