y0news
AnalyticsDigestsRSSAICrypto
#fp4-training1 article
1 articles
AIBullisharXiv โ€“ CS AI ยท 5h ago1
๐Ÿง 

Practical FP4 Training for Large-Scale MoE Models on Hopper GPUs

Researchers developed a training method for large-scale Mixture-of-Experts (MoE) models using FP4 precision on Hopper GPUs without native 4-bit support. The technique achieves 14.8% memory reduction and 12.5% throughput improvement for 671B parameter models by using FP4 for activations while keeping core computations in FP8.