511 articles tagged with #reinforcement-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv β CS AI Β· Mar 57/10
π§ Researchers propose SaFeR, a new AI system for generating safety-critical scenarios to test autonomous driving systems. The approach uses transformer-based models with a novel resampling strategy to balance adversarial testing, physical feasibility, and realistic behavior in autonomous vehicle simulations.
AIBullisharXiv β CS AI Β· Mar 56/10
π§ Researchers demonstrate that multi-agent competitive training enables AI agents to develop agile flight capabilities and strategic behaviors that outperform traditional single-agent training methods. The approach shows superior sim-to-real transfer and generalization when applied to drone racing scenarios with complex environments and obstacles.
AIBullisharXiv β CS AI Β· Mar 56/10
π§ Researchers developed R1-Code-Interpreter, a large language model that uses multi-stage reinforcement learning to autonomously generate code for step-by-step reasoning across diverse tasks. The 14B parameter model achieves 72.4% accuracy on test tasks, outperforming GPT-4o variants and demonstrating emergent self-checking capabilities through code generation.
π’ Hugging Faceπ§ GPT-4
AIBullisharXiv β CS AI Β· Mar 57/10
π§ Researchers have developed Sim2Sea, a comprehensive framework that successfully bridges the simulation-to-reality gap for autonomous maritime vessel navigation in congested waters. The system uses GPU-accelerated parallel simulation, dual-stream spatiotemporal policy, and targeted domain randomization to achieve zero-shot transfer from simulation to real-world deployment on a 17-ton unmanned vessel.
AIBullisharXiv β CS AI Β· Mar 57/10
π§ Researchers have developed a new framework for robotic agents that can adapt and learn continuously during operation, rather than being limited to fixed parameters from offline training. The system uses world model prediction residuals to detect unexpected events and automatically trigger self-improvement without external supervision.
AIBullisharXiv β CS AI Β· Mar 56/10
π§ Researchers introduce SHE (Stepwise Hybrid Examination), a new reinforcement learning framework that improves AI-powered e-commerce search relevance prediction. The framework addresses limitations in existing training methods by using step-level rewards and hybrid verification to enhance both accuracy and interpretability of search results.
AIBullisharXiv β CS AI Β· Mar 47/103
π§ Researchers have identified a critical flaw in reinforcement learning fine-tuning of large language models that causes degradation in multi-attempt performance despite improvements in single attempts. Their proposed solution, Diversity-Preserving Hybrid RL (DPH-RL), uses mass-covering f-divergences to maintain model diversity and prevent catastrophic forgetting while improving training efficiency.
AIBullisharXiv β CS AI Β· Mar 47/103
π§ Researchers introduce NE-Dreamer, a decoder-free model-based reinforcement learning agent that uses temporal transformers to predict next-step encoder embeddings. The approach achieves performance matching or exceeding DreamerV3 on standard benchmarks while showing substantial improvements on memory and spatial reasoning tasks.
AIBullisharXiv β CS AI Β· Mar 46/102
π§ Researchers identified a critical problem in Large Audio-Language Models (LALMs) where audio perception deteriorates during extended reasoning processes. They developed MPARΒ² framework using reinforcement learning, which improved perception performance from 31.74% to 63.51% and achieved 74.59% accuracy on MMAU benchmark.
AIBullisharXiv β CS AI Β· Mar 47/103
π§ Researchers developed ATPO (Adaptive Tree Policy Optimization), a new AI algorithm for multi-turn medical dialogues that outperforms existing methods by better handling uncertainty in patient-doctor interactions. The algorithm enabled a smaller Qwen3-8B model to surpass GPT-4o's accuracy by 0.92% on medical dialogue benchmarks through improved value estimation and exploration strategies.
AIBullisharXiv β CS AI Β· Mar 47/102
π§ Researchers have released MedXIAOHE, a new medical vision-language AI foundation model that achieves state-of-the-art performance across medical benchmarks and surpasses leading closed-source systems. The model incorporates advanced features like entity-aware pretraining, reinforcement learning for medical reasoning, and evidence-grounded report generation to improve reliability in clinical applications.
AINeutralarXiv β CS AI Β· Mar 47/104
π§ Researchers introduce GraphSSR, a new framework that improves zero-shot graph learning by combining Large Language Models with adaptive subgraph denoising. The system addresses structural noise issues in existing methods through a dynamic 'Sample-Select-Reason' pipeline and reinforcement learning training.
AIBullisharXiv β CS AI Β· Mar 46/102
π§ Researchers introduce Perception-R1, a new approach to enhance multimodal reasoning in large language models by improving visual perception capabilities through reinforcement learning with visual perception rewards. The method achieves state-of-the-art performance on multimodal reasoning benchmarks using only 1,442 training samples.
AIBullisharXiv β CS AI Β· Mar 46/103
π§ Researchers have developed TikZilla, a new AI model that generates high-quality scientific figures from text descriptions using TikZ code. The model uses a dataset four times larger than previous versions and combines supervised learning with reinforcement learning to achieve performance matching GPT-5 while using much smaller model sizes.
AIBullisharXiv β CS AI Β· Mar 46/103
π§ Researchers introduce RAPO (Retrieval-Augmented Policy Optimization), a new reinforcement learning framework that improves LLM agent training by incorporating retrieval mechanisms for broader exploration. The method achieves 5% performance gains across 14 datasets and 1.2x faster training efficiency by using hybrid-policy rollouts and retrieval-aware optimization.
AIBullisharXiv β CS AI Β· Mar 46/102
π§ Researchers propose NAR-CP, a new method to improve Large Language Models' performance in high-frequency decision-making tasks like UAV pursuit. The approach uses normalized action rewards and consistency policy optimization to address limitations in current LLM-based agents that struggle with rapid, precise numerical state updates.
AIBullisharXiv β CS AI Β· Mar 47/103
π§ Researchers introduce Skywork-Reward-V2, a suite of AI reward models trained on SynPref-40M, a massive 40-million preference pair dataset created through human-AI collaboration. The models achieve state-of-the-art performance across seven major benchmarks by combining human annotation quality with AI scalability for better preference learning.
AIBullisharXiv β CS AI Β· Mar 47/103
π§ Researchers propose Contextualized Defense Instructing (CDI), a new privacy defense paradigm for LLM agents that uses reinforcement learning to generate context-aware privacy guidance during execution. The approach achieves 94.2% privacy preservation while maintaining 80.6% helpfulness, outperforming static defense methods.
AINeutralarXiv β CS AI Β· Mar 47/103
π§ New research provides theoretical analysis of reinforcement learning's impact on Large Language Model planning capabilities, revealing that RL improves generalization through exploration while supervised fine-tuning may create spurious solutions. The study shows Q-learning maintains output diversity better than policy gradient methods, with findings validated on real-world planning benchmarks.
AIBullisharXiv β CS AI Β· Mar 46/103
π§ Researchers propose RL3DEdit, a reinforcement learning framework that addresses multi-view consistency challenges in 3D scene editing by using 2D diffusion model priors with novel reward signals from 3D foundation models. The method achieves stable multi-view consistency and outperforms existing approaches in editing quality and efficiency.
AIBullisharXiv β CS AI Β· Mar 46/103
π§ Researchers developed COOL-MC, a tool that combines reinforcement learning with model checking to verify and explain AI policies for platelet inventory management in blood banks. The system achieved a 2.9% stockout probability while providing transparent decision-making explanations for safety-critical healthcare applications.
AIBullisharXiv β CS AI Β· Mar 46/103
π§ Researchers introduce TΒ³, a new method to improve large language model (LLM) agents' reasoning abilities by tracking and correcting 'belief deviation' - when AI agents lose accurate understanding of problem states. The technique achieved up to 30-point performance gains and 34% token cost reduction across challenging tasks.
$COMP
AIBullisharXiv β CS AI Β· Mar 46/105
π§ Researchers introduce CORE (Concept-Oriented REinforcement), a new training framework that improves large language models' mathematical reasoning by bridging the gap between memorizing definitions and applying concepts. The method uses concept-aligned quizzes and concept-primed trajectories to provide fine-grained supervision, showing consistent improvements over traditional training approaches across multiple benchmarks.
AINeutralarXiv β CS AI Β· Mar 47/102
π§ Researchers propose the 'latent value hypothesis' to explain why Reinforcement Learning from AI Feedback (RLAIF) enables language models to self-improve through their own preference judgments. The theory suggests that pretraining on internet-scale data encodes human values in representation space, which constitutional prompts can elicit for value alignment.
AIBullisharXiv β CS AI Β· Mar 37/105
π§ Researchers have developed DeepMedix-R1, a foundation model for chest X-ray interpretation that provides transparent, step-by-step reasoning alongside accurate diagnoses to address the black-box problem in medical AI. The model uses reinforcement learning to align diagnostic outputs with clinical plausibility and significantly outperforms existing models in report generation and visual question answering tasks.