AINeutralarXiv โ CS AI ยท 3h ago6/10
๐ง
Rethinking Network Topologies for Cost-Effective Mixture-of-Experts LLM Serving
Researchers challenge the necessity of expensive high-bandwidth networks for Mixture-of-Experts LLM serving, demonstrating that lower-cost switchless topologies deliver 20.6-56.2% better cost-effectiveness than industry-standard scale-up architectures. The analysis reveals current network infrastructure is over-provisioned, with implications for data center economics and AI deployment efficiency.