AIBullisharXiv โ CS AI ยท 15h ago6/10
๐ง
VLMQ: Token Saliency-Driven Post-Training Quantization for Vision-language Models
Researchers introduced VLMQ, a post-training quantization framework specifically designed for vision-language models that addresses visual over-representation and modality gaps. The method achieves significant performance improvements, including 16.45% better results on MME-RealWorld under 2-bit quantization compared to existing approaches.