y0news
AnalyticsDigestsSourcesRSSAICrypto
#token-saliency1 article
1 articles
AIBullisharXiv โ€“ CS AI ยท 15h ago6/10
๐Ÿง 

VLMQ: Token Saliency-Driven Post-Training Quantization for Vision-language Models

Researchers introduced VLMQ, a post-training quantization framework specifically designed for vision-language models that addresses visual over-representation and modality gaps. The method achieves significant performance improvements, including 16.45% better results on MME-RealWorld under 2-bit quantization compared to existing approaches.