AINeutralarXiv – CS AI · 10h ago6/10
🧠
Fitting Is Not Enough: Smoothness in Extremely Quantized LLMs
Researchers demonstrate that extreme quantization of large language models causes degradation beyond numerical precision loss, specifically through reduced smoothness in prediction spaces. They introduce smoothness-preserving techniques in post-training and quantization-aware training that improve generation quality independent of numerical accuracy gains.