AIBullisharXiv โ CS AI ยท 7h ago7/10
๐ง
Cost-Efficient Multimodal LLM Inference via Cross-Tier GPU Heterogeneity
Researchers developed HeteroServe, a system that optimizes multimodal large language model inference by partitioning vision encoding and language generation across different GPU tiers. The approach reduces data transfer requirements and achieves 31-40% cost savings while improving throughput by up to 54% compared to existing systems.