y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#reinforcement-learning News & Analysis

511 articles tagged with #reinforcement-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

511 articles
AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

MagicAgent: Towards Generalized Agent Planning

Researchers have developed MagicAgent, a series of foundation models designed for generalized AI agent planning that outperforms existing sub-100B models and even surpasses leading ultra-scale models like GPT-5.2. The models achieve superior performance through a novel synthetic data framework and two-stage training paradigm that addresses gradient interference in multi-task learning.

AIBearisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

On The Fragility of Benchmark Contamination Detection in Reasoning Models

New research reveals that benchmark contamination in language reasoning models (LRMs) is extremely difficult to detect, allowing developers to easily inflate performance scores on public leaderboards. The study shows that reinforcement learning methods like GRPO and PPO can effectively conceal contamination signals, undermining the integrity of AI model evaluations.

$NEAR
AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

AReaL: A Large-Scale Asynchronous Reinforcement Learning System for Language Reasoning

Researchers have developed AReaL, a new asynchronous reinforcement learning system that dramatically improves the efficiency of training large language models for reasoning tasks. The system achieves up to 2.77x training speedup compared to traditional synchronous methods by decoupling generation from training processes.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

EnterpriseBench Corecraft: Training Generalizable Agents on High-Fidelity RL Environments

Surge AI introduces CoreCraft, the first environment in EnterpriseBench for training AI agents on realistic enterprise workflows. Training GLM 4.6 on this high-fidelity customer support simulation improved task performance from 25% to 37% and showed positive transfer to other benchmarks, demonstrating that quality training environments enable generalizable AI capabilities.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

AgentOCR: Reimagining Agent History via Optical Self-Compression

Researchers introduce AgentOCR, a framework that converts AI agent interaction histories from text to compressed visual format, reducing token usage by over 50% while maintaining 95% performance. The system uses visual caching and adaptive compression to address memory bottlenecks in large language model deployments.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

AceGRPO: Adaptive Curriculum Enhanced Group Relative Policy Optimization for Autonomous Machine Learning Engineering

Researchers introduce AceGRPO, a new reinforcement learning framework for Autonomous Machine Learning Engineering that addresses behavioral stagnation in current LLM-based agents. The Ace-30B model trained with this method achieves 100% valid submission rate on MLE-Bench-Lite and matches performance of proprietary frontier models while outperforming larger open-source alternatives.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

MAS-Orchestra: Understanding and Improving Multi-Agent Reasoning Through Holistic Orchestration and Controlled Benchmarks

Researchers introduce MAS-Orchestra, a new framework for multi-agent AI systems that uses reinforcement learning to orchestrate multiple AI agents more efficiently. The system achieves 10x efficiency improvements over existing methods and includes a benchmark (MASBENCH) to better understand when multi-agent systems outperform single-agent approaches.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

ZeroDVFS: Zero-Shot LLM-Guided Core and Frequency Allocation for Embedded Platforms

Researchers developed ZeroDVFS, a system that uses Large Language Models to optimize power management in embedded systems without requiring extensive profiling. The system achieves 7.09 times better energy efficiency and enables zero-shot deployment for new workloads in under 5 seconds through LLM-based code analysis.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

DRPO: Efficient Reasoning via Decoupled Reward Policy Optimization

Researchers propose Decoupled Reward Policy Optimization (DRPO), a new framework that reduces computational costs in large reasoning models by 77% while maintaining performance. The method addresses the 'overthinking' problem where AI models generate unnecessarily long reasoning for simple questions, achieving significant efficiency gains over existing approaches.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Stable Asynchrony: Variance-Controlled Off-Policy RL for LLMs

MIT researchers introduce VCPO (Variance Controlled Policy Optimization), a new method that improves asynchronous reinforcement learning for LLM training by addressing high variance issues in off-policy settings. The technique dynamically scales learning rates and applies variance control to achieve stable training with 2.5x speedup while maintaining performance.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

Value Flows

Researchers have developed Value Flows, a new reinforcement learning method that uses flow-based models to estimate complete return distributions rather than single scalar values. The approach achieves 1.3x improvement in success rates across 62 benchmark tasks by better identifying states with high return uncertainty for improved decision-making.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

VITA: Zero-Shot Value Functions via Test-Time Adaptation of Vision-Language Models

Researchers introduce VITA, a zero-shot value function learning method that enhances Vision-Language Models through test-time adaptation for robotic manipulation tasks. The system updates parameters sequentially over trajectories to improve temporal reasoning and generalizes across diverse environments, outperforming existing autoregressive VLM methods.

AIBullisharXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

Toward Clinically Explainable AI for Medical Diagnosis: A Foundation Model with Human-Compatible Reasoning via Reinforcement Learning

Researchers have developed DeepMedix-R1, a foundation model for chest X-ray interpretation that provides transparent, step-by-step reasoning alongside accurate diagnoses to address the black-box problem in medical AI. The model uses reinforcement learning to align diagnostic outputs with clinical plausibility and significantly outperforms existing models in report generation and visual question answering tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

GEM: A Gym for Agentic LLMs

Researchers introduced GEM (General Experience Maker), an open-source environment simulator designed for training large language models through experience-based learning rather than static datasets. The framework provides a standardized interface similar to OpenAI-Gym but specifically optimized for LLMs, featuring diverse environments, integrated tools, and compatibility with popular RL training frameworks.

$MKR
AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

GAR: Generative Adversarial Reinforcement Learning for Formal Theorem Proving

Researchers introduce GAR (Generative Adversarial Reinforcement Learning), a new AI training framework that jointly trains problem generators and solvers in an adversarial loop for formal theorem proving. The method shows significant improvements in mathematical proof capabilities, with models achieving 4.20% average relative improvement on benchmark tests.

AINeutralarXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Reasoning or Retrieval? A Study of Answer Attribution on Large Reasoning Models

Researchers discovered that large reasoning models (LRMs) suffer from inconsistent answers due to competing mechanisms between Chain-of-Thought reasoning and memory retrieval. They developed FARL, a new fine-tuning framework that suppresses retrieval shortcuts to promote genuine reasoning capabilities in AI models.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

CharacterFlywheel: Scaling Iterative Improvement of Engaging and Steerable LLMs in Production

Meta presents CharacterFlywheel, an iterative process for improving large language models in production social chat applications across Instagram, WhatsApp, and Messenger. Starting from LLaMA 3.1, the system achieved significant improvements through 15 generations of refinement, with the best models showing up to 8.8% improvement in engagement breadth and 19.4% in engagement depth while substantially improving instruction following capabilities.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Doctor-R1: Mastering Clinical Inquiry with Experiential Agentic Reinforcement Learning

Doctor-R1 is a new AI agent that combines accurate medical decision-making with strategic, empathetic patient consultation skills through reinforcement learning. The system outperforms existing open-source medical LLMs and proprietary models on clinical benchmarks while demonstrating superior communication quality and patient-centric performance.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

AgentMath: Empowering Mathematical Reasoning for Large Language Models via Tool-Augmented Agent

Researchers introduced AgentMath, a new AI framework that combines language models with code interpreters to solve complex mathematical problems more efficiently than current Large Reasoning Models. The system achieves state-of-the-art performance on mathematical competition benchmarks, with AgentMath-30B-A3B reaching 90.6% accuracy on AIME24 while remaining competitive with much larger models like OpenAI-o3.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Self-Harmony: Learning to Harmonize Self-Supervision and Self-Play in Test-Time Reinforcement Learning

Researchers introduce Self-Harmony, a new test-time reinforcement learning framework that improves AI model accuracy by having models solve problems and rephrase questions simultaneously. The method uses harmonic mean aggregation instead of majority voting to select stable answers, achieving state-of-the-art results across 28 of 30 reasoning benchmarks without requiring human supervision.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

RLP: Reinforcement as a Pretraining Objective

Researchers introduce RLP (Reinforcement Learning Pretraining), a new training method that incorporates reinforcement learning exploration into the pretraining phase rather than only post-training. The approach treats chain-of-thought reasoning as exploratory actions and achieved 19% performance improvements on math and science benchmarks across different model architectures.

$COMP
AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

Language Agents for Hypothesis-driven Clinical Decision Making with Reinforcement Learning

Researchers developed LA-CDM, a language agent that uses reinforcement learning to support clinical decision-making by iteratively requesting tests and generating hypotheses for diagnosis. The system was trained using a hybrid approach combining supervised and reinforcement learning, and tested on real-world data covering four abdominal diseases.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

LongWriter-Zero: Mastering Ultra-Long Text Generation via Reinforcement Learning

Researchers introduce LongWriter-Zero, a reinforcement learning approach that enables large language models to generate ultra-long, high-quality text without relying on synthetic training data. The 32B parameter model outperforms traditional supervised fine-tuning methods and even surpasses larger 100B+ models on long-form writing benchmarks.

AINeutralarXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Selection as Power: Constrained Reinforcement for Bounded Decision Authority

Researchers extend the "Selection as Power" framework to dynamic settings, introducing constrained reinforcement learning that maintains bounded decision authority in AI systems. The study demonstrates that governance constraints can prevent AI systems from collapsing into deterministic dominance while still allowing adaptive improvement through controlled parameter updates.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

Scaf-GRPO: Scaffolded Group Relative Policy Optimization for Enhancing LLM Reasoning

Researchers introduced Scaf-GRPO, a new training framework that overcomes the 'learning cliff' problem in LLM reasoning by providing strategic hints when models plateau. The method boosted Qwen2.5-Math-7B performance on the AIME24 benchmark by 44.3% relative to baseline GRPO methods.