y0news
AnalyticsDigestsSourcesRSSAICrypto
#low-precision-training1 article
1 articles
AIBullisharXiv โ€“ CS AI ยท 5d ago7/104
๐Ÿง 

A Convergence Analysis of Adaptive Optimizers under Floating-point Quantization

Researchers introduce the first theoretical framework analyzing convergence of adaptive optimizers like Adam and Muon under floating-point quantization in low-precision training. The study shows these algorithms maintain near full-precision performance when mantissa length scales logarithmically with iterations, with Muon proving more robust than Adam to quantization errors.