15 articles tagged with #agent-architecture. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท 2d ago7/10
๐ง Researchers introduce dual-trace memory encoding for LLM agents, pairing factual records with narrative scene reconstructions to improve cross-session recall by 20+ percentage points. The method significantly enhances temporal reasoning and multi-session knowledge aggregation without increasing computational costs, advancing the capability of persistent AI agent systems.
AINeutralarXiv โ CS AI ยท 3d ago7/10
๐ง Researchers introduce PAC-Bench, a benchmark for evaluating how AI agents collaborate while maintaining privacy constraints. The study reveals that privacy protections significantly degrade multi-agent system performance and identify coordination failures as a critical unsolved challenge requiring new technical approaches.
$PAC
AIBullisharXiv โ CS AI ยท Mar 267/10
๐ง Researchers have developed Declarative Model Interface (DMI), a new abstraction layer that transforms traditional GUIs into LLM-friendly interfaces for computer-use agents. Testing with Microsoft Office Suite showed 67% improvement in task success rates and 43.5% reduction in interaction steps, with over 61% of tasks completed in a single LLM call.
AIBullisharXiv โ CS AI ยท Mar 177/10
๐ง Researchers developed a two-agent defense system called OpenClaw that achieved 0% attack success rate against prompt injection attacks on LLM applications. The system uses agent isolation and JSON formatting to structurally prevent malicious prompts from reaching action-taking agents.
AIBullisharXiv โ CS AI ยท Mar 167/10
๐ง Researchers introduce the AI Search Paradigm, a comprehensive framework for next-generation search systems using four LLM-powered agents (Master, Planner, Executor, Writer) that collaborate to handle everything from simple queries to complex reasoning tasks. The system employs modular architecture with dynamic workflows for task planning, tool integration, and content synthesis to create more adaptive and scalable AI search capabilities.
AINeutralarXiv โ CS AI ยท Mar 46/103
๐ง Researchers prove 'selection theorems' showing that AI agents achieving low regret on prediction tasks must develop internal predictive models and belief states. The work demonstrates that structured internal representations are mathematically necessary, not just helpful, for competent decision-making under uncertainty.
AINeutralarXiv โ CS AI ยท 2d ago6/10
๐ง Researchers investigated whether self-monitoring mechanisms (metacognition, self-prediction, duration estimation) improve reinforcement learning agents in predator-prey environments. Initial auxiliary-loss implementations provided no benefits, but structurally integrating these modules into decision pathways showed modest improvements, suggesting effective AI enhancement requires architectural embedding rather than add-on approaches.
AINeutralarXiv โ CS AI ยท 2d ago6/10
๐ง Researchers demonstrate that large language models develop attractor-like geometric patterns in their activation space when processing identity documents describing persistent agents. Experiments on Llama 3.1 and Gemma 2 show paraphrased identity descriptions cluster significantly tighter than structural controls, suggesting LLMs encode semantic agent identity as stable attractors independent of linguistic variation.
๐ง Llama
AIBullisharXiv โ CS AI ยท 2d ago6/10
๐ง Researchers introduce Mโ
, a method that automatically evolves task-specific memory systems for large language model agents by treating memory architecture as executable Python code. The approach outperforms fixed memory designs across conversation, planning, and reasoning benchmarks, suggesting that specialized memory mechanisms significantly outperform one-size-fits-all solutions.
AINeutralarXiv โ CS AI ยท 3d ago6/10
๐ง Researchers propose SGH (Structured Graph Harness), a framework that replaces iterative Agent Loops with explicit directed acyclic graphs (DAGs) for LLM agent execution. The approach addresses structural weaknesses in current agent design by enforcing immutable execution plans, separating planning from recovery, and implementing strict escalation protocols, trading some flexibility for improved controllability and verifiability.
AINeutralarXiv โ CS AI ยท Apr 106/10
๐ง Researchers introduce a declarative runtime protocol that externalizes agent state to measure how much of an LLM-based agent's competence actually derives from the language model versus explicit structural components. Testing on Collaborative Battleship, they find that explicit world-model planning drives most performance gains, while sparse LLM-based revision at 4.3% of turns yields minimal and sometimes negative returns.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers introduce a formal planning framework that maps LLM-based web agents to traditional search algorithms, enabling better diagnosis of failures in autonomous web tasks. The study compares different agent architectures using novel evaluation metrics and a dataset of 794 human-labeled trajectories from WebArena benchmark.
AINeutralarXiv โ CS AI ยท Mar 96/10
๐ง Researchers have developed ESAA-Security, a new architecture for conducting secure, verifiable audits of AI-generated code using structured agent workflows rather than unstructured LLM conversations. The system creates an immutable audit trail through event-sourcing and produces comprehensive security reports across 26 tasks and 95 executable checks.
AIBullisharXiv โ CS AI ยท Mar 36/1012
๐ง Researchers developed Self-Healing Router, a fault-tolerant system for LLM agents that reduces control-plane LLM calls by 93% while maintaining correctness. The system uses graph-based routing with automatic recovery mechanisms, treating agent decisions as routing problems rather than reasoning tasks.
$COMP
AINeutralarXiv โ CS AI ยท Apr 74/10
๐ง Researchers developed a minimal AI architecture where a 'perspective latent' creates history-dependent perception in artificial agents. The system allows identical observations to be processed differently based on accumulated experience, demonstrating measurable plasticity that persists even after conditions return to normal.