y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#multi-agent News & Analysis

85 articles tagged with #multi-agent. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

85 articles
AIBullisharXiv – CS AI · Mar 56/10
🧠

Agile Flight Emerges from Multi-Agent Competitive Racing

Researchers demonstrate that multi-agent competitive training enables AI agents to develop agile flight capabilities and strategic behaviors that outperform traditional single-agent training methods. The approach shows superior sim-to-real transfer and generalization when applied to drone racing scenarios with complex environments and obstacles.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Learning Approximate Nash Equilibria in Cooperative Multi-Agent Reinforcement Learning via Mean-Field Subsampling

Researchers propose ALTERNATING-MARL, a new framework for cooperative multi-agent reinforcement learning that enables a global agent to learn with massive populations under communication constraints. The method achieves approximate Nash equilibrium convergence while only observing a subset of local agent states, with applications in multi-robot control and federated optimization.

$MKR
AIBullisharXiv – CS AI · Mar 56/10
🧠

MAGE: Meta-Reinforcement Learning for Language Agents toward Strategic Exploration and Exploitation

Researchers propose MAGE, a meta-reinforcement learning framework that enables Large Language Model agents to strategically explore and exploit in multi-agent environments. The framework uses multi-episode training with interaction histories and reflections, showing superior performance compared to existing baselines and strong generalization to unseen opponents.

AIBullisharXiv – CS AI · Mar 46/103
🧠

MA-CoNav: A Master-Slave Multi-Agent Framework with Hierarchical Collaboration and Dual-Level Reflection for Long-Horizon Embodied VLN

Researchers propose MA-CoNav, a multi-agent collaborative framework for robot navigation that uses a Master-Slave architecture to distribute cognitive tasks among specialized agents. The system outperforms existing Vision-Language Navigation methods by decoupling perception, planning, execution, and memory functions across different AI agents with hierarchical collaboration.

AIBullisharXiv – CS AI · Mar 46/106
🧠

SuperLocalMemory: Privacy-Preserving Multi-Agent Memory with Bayesian Trust Defense Against Memory Poisoning

SuperLocalMemory is a new privacy-preserving memory system for multi-agent AI that defends against memory poisoning attacks through local-first architecture and Bayesian trust scoring. The open-source system eliminates cloud dependencies while providing personalized retrieval through adaptive learning-to-rank, demonstrating strong performance metrics including 10.6ms search latency and 72% trust degradation for sleeper attacks.

AIBullisharXiv – CS AI · Mar 47/102
🧠

Saarthi for AGI: Towards Domain-Specific General Intelligence for Formal Verification

Researchers have enhanced the Saarthi AI framework for formal verification, achieving 70% better accuracy in generating SystemVerilog assertions and 50% fewer iterations to reach coverage closure. The framework uses multi-agent collaboration and improved RAG techniques to move toward domain-specific AI intelligence for verification tasks.

AIBullisharXiv – CS AI · Mar 46/102
🧠

AI-for-Science Low-code Platform with Bayesian Adversarial Multi-Agent Framework

Researchers have developed a Bayesian adversarial multi-agent framework for AI-driven scientific code generation, featuring three coordinated LLM agents that work together to improve reliability and reduce errors. The Low-code Platform (LCP) enables non-expert users to generate scientific code through natural language prompts, demonstrating superior performance in benchmark tests and Earth Science applications.

AIBullisharXiv – CS AI · Mar 47/102
🧠

ShareVerse: Multi-Agent Consistent Video Generation for Shared World Modeling

ShareVerse is a new AI video generation framework that enables multiple agents to interact and generate consistent videos within a shared virtual world. The system uses CARLA simulation data and cross-agent attention mechanisms to create 49-frame videos with multi-view consistency across different agents.

AIBullisharXiv – CS AI · Mar 47/103
🧠

BrandFusion: A Multi-Agent Framework for Seamless Brand Integration in Text-to-Video Generation

Researchers introduce BrandFusion, a multi-agent AI framework that enables seamless brand integration into text-to-video generation models. The system addresses commercial monetization challenges in T2V technology by automatically embedding advertiser brands into generated videos while preserving user intent and ensuring natural integration.

AIBullisharXiv – CS AI · Mar 46/102
🧠

Multimodal Multi-Agent Ransomware Analysis Using AutoGen

Researchers developed a multimodal multi-agent ransomware analysis framework using AutoGen that combines static, dynamic, and network data sources for improved ransomware detection. The system achieved 0.936 Macro-F1 score for family classification and demonstrated stable convergence over 100 epochs with a final composite score of 0.88.

AIBullisharXiv – CS AI · Mar 37/105
🧠

Elo-Evolve: A Co-evolutionary Framework for Language Model Alignment

Researchers introduce Elo-Evolve, a new framework for training AI language models using dynamic multi-agent competition instead of static reward functions. The method achieves 4.5x noise reduction and demonstrates superior performance compared to traditional alignment approaches when tested on Qwen2.5-7B models.

AIBullisharXiv – CS AI · Feb 277/105
🧠

CourtGuard: A Model-Agnostic Framework for Zero-Shot Policy Adaptation in LLM Safety

Researchers introduce CourtGuard, a new framework for AI safety that uses retrieval-augmented multi-agent debate to evaluate LLM outputs without requiring expensive retraining. The system achieves state-of-the-art performance across 7 safety benchmarks and demonstrates zero-shot adaptability to new policy requirements, offering a more flexible approach to AI governance.

AINeutralGoogle Research Blog · Jan 287/106
🧠

Towards a science of scaling agent systems: When and why agent systems work

The article discusses the scientific principles behind scaling agent systems in generative AI, examining the conditions and factors that determine when agent systems perform effectively. It appears to focus on understanding the theoretical foundations for building and deploying AI agent systems at scale.

AIBullishOpenAI News · Oct 237/105
🧠

Consensus accelerates research with GPT-5 and Responses API

Consensus has deployed GPT-5 and OpenAI's Responses API to create a multi-agent research assistant that can rapidly read, analyze, and synthesize scientific evidence. The platform serves over 8 million researchers and aims to accelerate scientific discovery by automating research processes that previously took much longer.

AIBullishOpenAI News · Mar 167/104
🧠

Learning to communicate

OpenAI has published new research demonstrating that AI agents can develop their own communication language. This research explores emergent communication capabilities in artificial intelligence systems.

AINeutralarXiv – CS AI · Apr 76/10
🧠

Implementing surrogate goals for safer bargaining in LLM-based agents

Researchers developed methods to implement 'surrogate goals' in LLM-based agents to reduce bargaining risks by deflecting threats away from what principals care about. The study tested four approaches (prompting, fine-tuning, scaffolding) and found that scaffolding and fine-tuning methods outperformed simple prompting for implementing desired threat response behaviors.

AINeutralarXiv – CS AI · Mar 176/10
🧠

InterveneBench: Benchmarking LLMs for Intervention Reasoning and Causal Study Design in Real Social Systems

Researchers introduced InterveneBench, a new benchmark comprising 744 peer-reviewed studies to evaluate large language models' ability to reason about policy interventions and causal inference in social science contexts. Current state-of-the-art LLMs struggle with this type of reasoning, prompting the development of STRIDES, a multi-agent framework that significantly improves performance on these tasks.

AIBullisharXiv – CS AI · Mar 176/10
🧠

EvolvR: Self-Evolving Pairwise Reasoning for Story Evaluation to Enhance Generation

Researchers have developed EvolvR, a self-evolving framework that improves AI's ability to evaluate and generate stories through pairwise reasoning and multi-agent data filtering. The system achieves state-of-the-art performance on three evaluation benchmarks and significantly enhances story generation quality when used as a reward model.

AINeutralarXiv – CS AI · Mar 176/10
🧠

More Agents Improve Math Problem Solving but Adversarial Robustness Gap Persists

Research reveals that while increasing the number of LLM agents improves mathematical problem-solving accuracy, these multi-agent systems remain vulnerable to adversarial attacks. The study found that human-like typos pose the greatest threat to robustness, and the adversarial vulnerability gap persists regardless of agent count.

🧠 Llama
AINeutralarXiv – CS AI · Mar 166/10
🧠

LLM Constitutional Multi-Agent Governance

Researchers introduce Constitutional Multi-Agent Governance (CMAG), a framework that prevents AI manipulation in multi-agent systems while maintaining cooperation. The study shows that unconstrained AI optimization achieves high cooperation but erodes agent autonomy and fairness, while CMAG preserves ethical outcomes with only modest cooperation reduction.

AIBullisharXiv – CS AI · Mar 126/10
🧠

Learning to Negotiate: Multi-Agent Deliberation for Collective Value Alignment in LLMs

Researchers propose a multi-agent negotiation framework for aligning large language models in scenarios involving conflicting stakeholder values. The approach uses two LLM instances with opposing personas engaging in structured dialogue to develop conflict resolution capabilities while maintaining collective agency alignment.

AIBullisharXiv – CS AI · Mar 116/10
🧠

LDP: An Identity-Aware Protocol for Multi-Agent LLM Systems

Researchers present LLM Delegate Protocol (LDP), a new AI-native communication protocol for multi-agent LLM systems that introduces identity awareness, progressive payloads, and governance mechanisms. The protocol achieves 12x lower latency on simple tasks and 37% token reduction compared to existing protocols like A2A, though quality improvements remain limited in small delegate pools.

← PrevPage 2 of 4Next →