AIBullisharXiv โ CS AI ยท 5d ago7/104
๐ง
A Convergence Analysis of Adaptive Optimizers under Floating-point Quantization
Researchers introduce the first theoretical framework analyzing convergence of adaptive optimizers like Adam and Muon under floating-point quantization in low-precision training. The study shows these algorithms maintain near full-precision performance when mantissa length scales logarithmically with iterations, with Muon proving more robust than Adam to quantization errors.