AINeutralarXiv โ CS AI ยท 10h ago6/10
๐ง
Provable Post-Training Quantization: Theoretical Analysis of OPTQ and Qronos
Researchers provide the first rigorous theoretical analysis of OPTQ (GPTQ), a widely-used post-training quantization algorithm for neural networks and LLMs, establishing quantitative error bounds and validating practical design choices. The study extends theoretical guarantees to both deterministic and stochastic variants of OPTQ and the Qronos algorithm, offering guidance for regularization parameter selection and quantization alphabet sizing.