y0news
AnalyticsDigestsSourcesRSSAICrypto
#low-bit1 article
1 articles
AIBullisharXiv โ€“ CS AI ยท Feb 276/106
๐Ÿง 

Q$^2$: Quantization-Aware Gradient Balancing and Attention Alignment for Low-Bit Quantization

Researchers propose Qยฒ, a new framework that addresses gradient imbalance issues in quantization-aware training for complex visual tasks like object detection and image segmentation. The method achieves significant performance improvements (+2.5% mAP for object detection, +3.7% mDICE for segmentation) while introducing no inference-time overhead.

$ADA