y0news
#ai-efficiency5 articles
5 articles
AIBullisharXiv โ€“ CS AI ยท 4h ago3
๐Ÿง 

RUMAD: Reinforcement-Unifying Multi-Agent Debate

Researchers introduce RUMAD, a reinforcement learning framework that optimizes multi-agent AI debate systems by dynamically controlling communication topology. The system achieves over 80% reduction in computational costs while improving reasoning accuracy across benchmark tests, with strong generalization capabilities across different task domains.

AIBullisharXiv โ€“ CS AI ยท 4h ago4
๐Ÿง 

FineScope : SAE-guided Data Selection Enables Domain Specific LLM Pruning and Finetuning

Researchers introduce FineScope, a framework that uses Sparse Autoencoder (SAE) techniques to create smaller, domain-specific language models from larger pretrained LLMs through structured pruning and self-data distillation. The method achieves competitive performance while significantly reducing computational requirements compared to training from scratch.

AIBullisharXiv โ€“ CS AI ยท 4h ago10
๐Ÿง 

MobileLLM-R1: Exploring the Limits of Sub-Billion Language Model Reasoners with Open Training Recipes

Researchers developed MobileLLM-R1, a sub-billion parameter AI model that demonstrates strong reasoning capabilities using only 2T tokens of high-quality data instead of massive 10T+ token datasets. The 950M parameter model achieves superior performance on reasoning benchmarks compared to larger competitors while using only 11.7% of the training data compared to proprietary models like Qwen3.

AIBullisharXiv โ€“ CS AI ยท 4h ago5
๐Ÿง 

DiffuMamba: High-Throughput Diffusion LMs with Mamba Backbone

Researchers introduce DiffuMamba, a new diffusion language model using Mamba backbone architecture that achieves up to 8.2x higher inference throughput than Transformer-based models while maintaining comparable performance. The model demonstrates linear scaling with sequence length and represents a significant advancement in efficient AI text generation systems.