WAPS-Quant: Low-Bit Post-Training Quantization Using Weight-Activation Product Scaling

Post-Training Quantization (PTQ) has been effectively compressing neural networks into very few bits using a limited calibration dataset. Various quantization methods utilizing second-order error have been proposed and demonstrated good performance. However, at extremely low bits, the increase in qu...

Full description

Saved in:
Bibliographic Details
Main Authors: Geunjae Choi, Kamin Lee, Nojun Kwak
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10982219/
Tags: Add Tag
No Tags, Be the first to tag this record!

Similar Items