AIBullisharXiv โ CS AI ยท 5h ago1
๐ง
Practical FP4 Training for Large-Scale MoE Models on Hopper GPUs
Researchers developed a training method for large-scale Mixture-of-Experts (MoE) models using FP4 precision on Hopper GPUs without native 4-bit support. The technique achieves 14.8% memory reduction and 12.5% throughput improvement for 671B parameter models by using FP4 for activations while keeping core computations in FP8.