y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#reinforcement-learning News & Analysis

511 articles tagged with #reinforcement-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

511 articles
AIBullishOpenAI News · Apr 277/105
🧠

OpenAI Gym Beta

OpenAI has released the public beta of OpenAI Gym, a comprehensive toolkit designed for developing and comparing reinforcement learning algorithms. The platform includes a diverse suite of environments ranging from simulated robots to Atari games, along with a website for result comparison and reproducibility.

AIBullisharXiv – CS AI · 14h ago6/10
🧠

Revisiting Entropy Regularization: Adaptive Coefficient Unlocks Its Potential for LLM Reinforcement Learning

Researchers propose Adaptive Entropy Regularization (AER), a dynamic framework that addresses policy entropy collapse in LLM reinforcement learning by adjusting exploration intensity based on task difficulty. The method improves upon fixed entropy regularization approaches, demonstrating consistent gains in mathematical reasoning benchmarks while maintaining balanced exploration-exploitation tradeoffs.

AINeutralarXiv – CS AI · 14h ago6/10
🧠

AtManRL: Towards Faithful Reasoning via Differentiable Attention Saliency

Researchers introduce AtManRL, a method that combines differentiable attention manipulation with reinforcement learning to improve the faithfulness of chain-of-thought reasoning in large language models. By training attention masks to identify which tokens genuinely influence model predictions, the approach demonstrates that LLM reasoning traces can be made more interpretable and transparent.

🧠 Llama
AINeutralarXiv – CS AI · 14h ago6/10
🧠

Reward Weighted Classifier-Free Guidance as Policy Improvement in Autoregressive Models

Researchers demonstrate that reward-weighted classifier-free guidance (RCFG) can dynamically adjust autoregressive model outputs to optimize arbitrary reward functions at test time without retraining. Applied to molecular generation, this approach enables real-time optimization of competing objectives and accelerates reinforcement learning convergence when used as a teacher for policy distillation.

AINeutralarXiv – CS AI · 14h ago6/10
🧠

Dynamic Sampling that Adapts: Self-Aware Iterative Data Persistent Optimization for Mathematical Reasoning

Researchers introduce SAI-DPO, a dynamic data sampling framework that adapts training data selection based on a model's evolving capabilities during training, rather than using static metrics. Tested on mathematical reasoning benchmarks including AIME24 and AMC23, SAI-DPO achieves state-of-the-art performance with significantly less training data, outperforming baselines by nearly 6 points.

AIBullisharXiv – CS AI · 14h ago6/10
🧠

EnvScaler: Scaling Tool-Interactive Environments for LLM Agent via Programmatic Synthesis

EnvScaler is an automated framework that generates synthetic tool-interaction environments for training LLM agents through programmatic synthesis, creating 191 diverse environments and 7,000 scenarios. The approach addresses scalability challenges in LLM agent training by combining topic mining and logic modeling to overcome hallucinations and manual bottlenecks, demonstrating improved performance on multi-turn, multi-tool interaction tasks.

AINeutralarXiv – CS AI · 14h ago6/10
🧠

Deliberative Searcher: Improving LLM Reliability via Reinforcement Learning with constraints

Researchers present Deliberative Searcher, a framework that enhances large language model reliability by combining certainty calibration with retrieval-based search for question answering. The system uses reinforcement learning with soft reliability constraints to improve alignment between model confidence and actual correctness, producing more trustworthy outputs.

AIBullisharXiv – CS AI · 14h ago6/10
🧠

"Excuse me, may I say something..." CoLabScience, A Proactive AI Assistant for Biomedical Discovery and LLM-Expert Collaborations

Researchers introduce CoLabScience, a proactive AI assistant designed to enhance biomedical research collaboration by intervening in scientific discussions at optimal moments. The system uses PULI, a reinforcement learning framework that learns when and how to contribute based on project context and conversation history, supported by a new benchmark dataset (BSDD) of simulated research dialogues.

AINeutralarXiv – CS AI · 5d ago6/10
🧠

Beyond Static Sandboxing: Learned Capability Governance for Autonomous AI Agents

Researchers introduce Aethelgard, an adaptive governance framework that addresses the capability overprovisioning problem in autonomous AI agents by dynamically restricting tool access based on task requirements. The system uses reinforcement learning to enforce least-privilege principles, reducing security exposure while maintaining operational efficiency.

AIBullisharXiv – CS AI · 5d ago6/10
🧠

KG-Reasoner: A Reinforced Model for End-to-End Multi-Hop Knowledge Graph Reasoning

Researchers introduce KG-Reasoner, an end-to-end framework that uses reinforcement learning to train large language models to perform multi-hop reasoning over knowledge graphs without decomposing tasks into isolated pipeline steps. The approach demonstrates competitive or superior performance across eight reasoning benchmarks by enabling LLMs to dynamically explore reasoning paths and backtrack when necessary.

AIBullisharXiv – CS AI · 5d ago6/10
🧠

PromptEcho: Annotation-Free Reward from Vision-Language Models for Text-to-Image Reinforcement Learning

Researchers introduce PromptEcho, a novel reward construction method for improving text-to-image model training that requires no human annotation or model fine-tuning. By leveraging frozen vision-language models to compute token-level alignment scores, the approach achieves significant performance gains on multiple benchmarks while remaining computationally efficient.

AINeutralarXiv – CS AI · 5d ago6/10
🧠

No More Stale Feedback: Co-Evolving Critics for Open-World Agent Learning

Researchers introduce ECHO, a reinforcement learning framework that co-evolves policy and critic models to address the problem of stale feedback in LLM agent training. The system uses cascaded rollouts and saturation-aware gain shaping to maintain synchronized, relevant critique as the agent's behavior improves over time, demonstrating enhanced stability and success rates in complex environments.

AINeutralarXiv – CS AI · 5d ago6/10
🧠

StableSketcher: Enhancing Diffusion Model for Pixel-based Sketch Generation via Visual Question Answering Feedback

StableSketcher is a novel AI framework that enhances diffusion models for generating pixel-based hand-drawn sketches with improved prompt fidelity. The approach combines fine-tuned variational autoencoders with a reinforcement learning reward function based on visual question answering, alongside a new SketchDUO dataset of instance-level sketches paired with captions and Q&A pairs.

🧠 Stable Diffusion
AINeutralarXiv – CS AI · 5d ago6/10
🧠

Reasoning about Intent for Ambiguous Requests

Researchers propose a method for large language models to handle ambiguous user requests by generating structured responses that enumerate multiple valid interpretations with corresponding answers, trained via reinforcement learning with dual reward objectives for coverage and precision.

AINeutralarXiv – CS AI · 5d ago6/10
🧠

Self-Monitoring Benefits from Structural Integration: Lessons from Metacognition in Continuous-Time Multi-Timescale Agents

Researchers investigated whether self-monitoring mechanisms (metacognition, self-prediction, duration estimation) improve reinforcement learning agents in predator-prey environments. Initial auxiliary-loss implementations provided no benefits, but structurally integrating these modules into decision pathways showed modest improvements, suggesting effective AI enhancement requires architectural embedding rather than add-on approaches.

AINeutralarXiv – CS AI · 5d ago6/10
🧠

Artificial Intelligence for Modeling and Simulation of Mixed Automated and Human Traffic

A comprehensive survey examines AI methodologies for simulating mixed autonomous and human-driven traffic, addressing critical gaps in current simulation tools. The research proposes a unified taxonomy of AI methods spanning agent-level behavior models, environment-level simulations, and physics-informed approaches to improve autonomous vehicle testing and validation.

AIBullisharXiv – CS AI · 5d ago6/10
🧠

Cycle-Consistent Search: Question Reconstructability as a Proxy Reward for Search Agent Training

Researchers propose Cycle-Consistent Search (CCS), a novel framework for training search agents using reinforcement learning without requiring gold-standard labeled data. The method leverages question reconstructability as a reward signal, using information bottlenecks to ensure agents learn from genuine search quality rather than surface-level linguistic patterns.

AINeutralarXiv – CS AI · 6d ago6/10
🧠

A Comparative Theoretical Analysis of Entropy Control Methods in Reinforcement Learning

Researchers present a theoretical framework comparing entropy control methods in reinforcement learning for LLMs, showing that covariance-based regularization outperforms traditional entropy regularization by avoiding policy bias and achieving asymptotic unbiasedness. This analysis addresses a critical scaling challenge in RL-based LLM training where rapid policy entropy collapse limits model performance.

AINeutralarXiv – CS AI · 6d ago6/10
🧠

A Queueing-Theoretic Framework for Dynamic Attack Surfaces: Data-Integrated Risk Analysis and Adaptive Defense

Researchers develop a queueing-theoretic framework that models cyber-attack surfaces as dynamic systems where vulnerabilities arrive and depart over time. Using reinforcement learning and Markov decision processes, they demonstrate an adaptive defense strategy that reduces active vulnerabilities by over 90% in software supply chains without increasing maintenance budgets.

AIBullisharXiv – CS AI · 6d ago6/10
🧠

Skill-SD: Skill-Conditioned Self-Distillation for Multi-turn LLM Agents

Researchers introduce Skill-SD, a novel training framework for multi-turn LLM agents that improves sample efficiency by converting successful agent trajectories into dynamic natural language skills that condition a teacher model. The approach combines reinforcement learning with self-distillation and achieves significant performance improvements over baseline methods on benchmark tasks.

AIBullisharXiv – CS AI · 6d ago6/10
🧠

TInR: Exploring Tool-Internalized Reasoning in Large Language Models

Researchers propose Tool-Internalized Reasoning (TInR), a framework that embeds tool knowledge directly into Large Language Models rather than relying on external tool documentation during reasoning. The TInR-U model uses a three-phase training pipeline combining knowledge alignment, supervised fine-tuning, and reinforcement learning to improve reasoning efficiency and performance across various tasks.

AIBullisharXiv – CS AI · 6d ago6/10
🧠

MMR-AD: A Large-Scale Multimodal Dataset for Benchmarking General Anomaly Detection with Multimodal Large Language Models

Researchers introduced MMR-AD, a large-scale multimodal dataset designed to benchmark general anomaly detection using Multimodal Large Language Models (MLLMs). The study reveals that current state-of-the-art MLLMs fall short of industrial requirements for anomaly detection, though a proposed baseline model called Anomaly-R1 demonstrates significant improvements through reasoning-based approaches enhanced by reinforcement learning.

← PrevPage 8 of 21Next →