y0news
#transformer-models4 articles
4 articles
AIBullisharXiv โ€“ CS AI ยท 6h ago2
๐Ÿง 

Efficient Long-Horizon GUI Agents via Training-Free KV Cache Compression

Researchers developed ST-Lite, a training-free KV cache compression framework that accelerates GUI agents by 2.45x while using only 10-20% of the cache budget. The solution addresses memory and latency constraints in Vision-Language Models for autonomous GUI interactions through specialized attention pattern optimization.

AIBearisharXiv โ€“ CS AI ยท 6h ago2
๐Ÿง 

CaptionFool: Universal Image Captioning Model Attacks

Researchers have developed CaptionFool, a universal adversarial attack that can manipulate AI image captioning models by modifying just 1.2% of image patches. The attack achieves 94-96% success rates in forcing models to generate arbitrary captions, including offensive content that can bypass content moderation systems.

AINeutralarXiv โ€“ CS AI ยท 6h ago1
๐Ÿง 

Theoretical Perspectives on Data Quality and Synergistic Effects in Pre- and Post-Training Reasoning Models

New theoretical research analyzes how Large Language Models learn during pretraining versus post-training phases, revealing that balanced pretraining data creates latent capabilities activated later, while supervised fine-tuning works best on small, challenging datasets and reinforcement learning requires large-scale data that isn't overly difficult.

AIBullisharXiv โ€“ CS AI ยท 6h ago0
๐Ÿง 

Noise reduction in BERT NER models for clinical entity extraction

Researchers developed a Noise Removal model to improve precision in clinical entity extraction using BERT-based Named Entity Recognition systems. The model uses advanced features like Probability Density Maps to identify weak vs strong predictions, reducing false positives by 50-90% in clinical NER applications.