y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#memory-systems News & Analysis

35 articles tagged with #memory-systems. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

35 articles
AIBearisharXiv โ€“ CS AI ยท 3d ago7/10
๐Ÿง 

ADAM: A Systematic Data Extraction Attack on Agent Memory via Adaptive Querying

Researchers have developed ADAM, a novel privacy attack that exploits vulnerabilities in Large Language Model agents' memory systems through adaptive querying, achieving up to 100% success rates in extracting sensitive information. The attack highlights critical security gaps in modern LLM-based systems that rely on memory modules and retrieval-augmented generation, underscoring the urgent need for privacy-preserving safeguards.

AINeutralarXiv โ€“ CS AI ยท Apr 107/10
๐Ÿง 

ATANT: An Evaluation Framework for AI Continuity

Researchers introduce ATANT, an open evaluation framework designed to measure whether AI systems can maintain coherent context and continuity across time without confusing information across different narratives. The framework achieves up to 100% accuracy in isolated scenarios but drops to 96% when managing 250 simultaneous narratives, revealing practical limitations in current AI memory architectures.

AIBullisharXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

MemMachine: A Ground-Truth-Preserving Memory System for Personalized AI Agents

MemMachine is an open-source memory system for AI agents that preserves conversational ground truth and achieves superior accuracy-efficiency tradeoffs compared to existing solutions. The system integrates short-term, long-term episodic, and profile memory while using 80% fewer input tokens than comparable systems like Mem0.

๐Ÿง  GPT-4๐Ÿง  GPT-5
AIBullisharXiv โ€“ CS AI ยท Apr 67/10
๐Ÿง 

Opal: Private Memory for Personal AI

Researchers present Opal, a private memory system for personal AI that uses trusted hardware enclaves and oblivious RAM to protect user data privacy while maintaining query accuracy. The system achieves 13 percentage point improvement in retrieval accuracy over semantic search and 29x higher throughput with 15x lower costs than secure baselines.

AINeutralarXiv โ€“ CS AI ยท Mar 127/10
๐Ÿง 

Multi-Agent Memory from a Computer Architecture Perspective: Visions and Challenges Ahead

Researchers propose treating multi-agent AI memory as a computer architecture problem, introducing a three-layer memory hierarchy and identifying critical protocol gaps. The paper highlights multi-agent memory consistency as the most pressing challenge for building scalable collaborative AI systems.

AINeutralarXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

LifeBench: A Benchmark for Long-Horizon Multi-Source Memory

Researchers introduce LifeBench, a new AI benchmark that tests long-term memory systems by requiring integration of both declarative and non-declarative memory across extended timeframes. Current state-of-the-art memory systems achieve only 55.2% accuracy on this challenging benchmark, highlighting significant gaps in AI's ability to handle complex, multi-source memory tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

ELMUR: External Layer Memory with Update/Rewrite for Long-Horizon RL Problems

Researchers developed ELMUR, a new AI architecture that uses external memory to help robots make better decisions over extremely long time periods. The system achieved 100% success on tasks requiring memory of up to one million steps and nearly doubled performance on robotic manipulation tasks compared to existing methods.

AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

AriadneMem: Threading the Maze of Lifelong Memory for LLM Agents

Researchers have developed AriadneMem, a new memory system for long-horizon LLM agents that addresses challenges in maintaining accurate memory under fixed context budgets. The system uses a two-phase pipeline with entropy-aware gating and conflict-aware coarsening to improve multi-hop reasoning while reducing runtime by 77.8% and using only 497 context tokens.

๐Ÿง  GPT-4
AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

PlugMem: A Task-Agnostic Plugin Memory Module for LLM Agents

Researchers propose PlugMem, a task-agnostic plugin memory module for LLM agents that structures episodic memories into knowledge-centric graphs for efficient retrieval. The system consistently outperforms existing memory designs across multiple benchmarks while maintaining transferability between different tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 46/106
๐Ÿง 

SuperLocalMemory: Privacy-Preserving Multi-Agent Memory with Bayesian Trust Defense Against Memory Poisoning

SuperLocalMemory is a new privacy-preserving memory system for multi-agent AI that defends against memory poisoning attacks through local-first architecture and Bayesian trust scoring. The open-source system eliminates cloud dependencies while providing personalized retrieval through adaptive learning-to-rank, demonstrating strong performance metrics including 10.6ms search latency and 72% trust degradation for sleeper attacks.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

LightMem: Lightweight and Efficient Memory-Augmented Generation

Researchers introduce LightMem, a new memory system for Large Language Models that mimics human memory structure with three stages: sensory, short-term, and long-term memory. The system achieves up to 7.7% better QA accuracy while reducing token usage by up to 106x and API calls by up to 159x compared to existing methods.

AINeutralarXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

Agentic Unlearning: When LLM Agent Meets Machine Unlearning

Researchers introduce 'agentic unlearning' through Synchronized Backflow Unlearning (SBU), a framework that removes sensitive information from both AI model parameters and persistent memory systems. The method addresses critical gaps in existing unlearning techniques by preventing cross-pathway recontamination between memory and parameters.

AIBullisharXiv โ€“ CS AI ยท Feb 277/105
๐Ÿง 

Towards Autonomous Memory Agents

Researchers introduce U-Mem, an autonomous memory agent system that actively acquires and validates knowledge for large language models. The system uses cost-aware knowledge extraction and semantic Thompson sampling to improve performance, showing significant gains on benchmarks like HotpotQA and AIME25.

AIBullisharXiv โ€“ CS AI ยท Feb 277/107
๐Ÿง 

LLMServingSim 2.0: A Unified Simulator for Heterogeneous and Disaggregated LLM Serving Infrastructure

Researchers have released LLMServingSim 2.0, a unified simulator that models the complex interactions between heterogeneous hardware and disaggregated software in large language model serving infrastructures. The simulator achieves 0.97% average error compared to real deployments while maintaining 10-minute simulation times for complex configurations.

$NEAR
AIBullisharXiv โ€“ CS AI ยท 2d ago6/10
๐Ÿง 

M$^\star$: Every Task Deserves Its Own Memory Harness

Researchers introduce Mโ˜…, a method that automatically evolves task-specific memory systems for large language model agents by treating memory architecture as executable Python code. The approach outperforms fixed memory designs across conversation, planning, and reasoning benchmarks, suggesting that specialized memory mechanisms significantly outperform one-size-fits-all solutions.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

The Missing Knowledge Layer in Cognitive Architectures for AI Agents

Researchers identify a critical architectural gap in leading AI agent frameworks (CoALA and JEPA), which lack an explicit Knowledge layer with distinct persistence semantics. The paper proposes a four-layer decomposition model with fundamentally different update mechanics for knowledge, memory, wisdom, and intelligence, with working implementations demonstrating feasibility.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

MERMAID: Memory-Enhanced Retrieval and Reasoning with Multi-Agent Iterative Knowledge Grounding for Veracity Assessment

Researchers introduce MERMAID, a memory-enhanced multi-agent framework for automated fact-checking that couples evidence retrieval with reasoning processes. The system achieves state-of-the-art performance on multiple benchmarks by reusing retrieved evidence across claims, reducing redundant searches and improving verification efficiency.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

If an LLM Were a Character, Would It Know Its Own Story? Evaluating Lifelong Learning in LLMs

Researchers introduce LIFESTATE-BENCH, a benchmark for evaluating lifelong learning capabilities in large language models through multi-turn interactions using narrative datasets like Hamlet. Testing shows nonparametric approaches significantly outperform parametric methods, but all models struggle with catastrophic forgetting over extended interactions, revealing fundamental limitations in LLM memory and consistency.

๐Ÿง  GPT-4๐Ÿง  Llama
AIBullisharXiv โ€“ CS AI ยท Apr 76/10
๐Ÿง 

SuperLocalMemory V3.3: The Living Brain -- Biologically-Inspired Forgetting, Cognitive Quantization, and Multi-Channel Retrieval for Zero-LLM Agent Memory Systems

Researchers have released SuperLocalMemory V3.3, an open-source AI agent memory system that operates entirely locally without cloud LLMs, implementing biologically-inspired forgetting mechanisms and multi-channel retrieval. The system achieves 70.4% performance on LoCoMo benchmarks while running on CPU only, addressing the paradox of AI agents having vast knowledge but poor conversational memory.

AIBullisharXiv โ€“ CS AI ยท Apr 76/10
๐Ÿง 

Memory Intelligence Agent

Researchers have developed Memory Intelligence Agent (MIA), a new AI framework that improves deep research agents through a Manager-Planner-Executor architecture with advanced memory systems. The framework enables continuous learning during inference and demonstrates superior performance across eleven benchmarks through enhanced cooperation between parametric and non-parametric memory systems.

AIBullisharXiv โ€“ CS AI ยท Mar 276/10
๐Ÿง 

ElephantBroker: A Knowledge-Grounded Cognitive Runtime for Trustworthy AI Agents

Researchers have introduced ElephantBroker, an open-source cognitive runtime system that combines knowledge graphs with vector storage to create more trustworthy AI agents with verifiable memory. The system implements comprehensive safety measures, evidence verification, and multi-organizational access controls for enterprise AI deployments.

AINeutralarXiv โ€“ CS AI ยท Mar 266/10
๐Ÿง 

Enhanced Mycelium of Thought (EMoT): A Bio-Inspired Hierarchical Reasoning Architecture with Strategic Dormancy and Mnemonic Encoding

Researchers introduced Enhanced Mycelium of Thought (EMoT), a bio-inspired AI reasoning framework that organizes cognitive processing into four hierarchical levels with strategic dormancy and memory encoding. The system achieved near-parity with Chain-of-Thought reasoning on complex problems but significantly underperformed on simple tasks, with 33-fold higher computational costs.

Page 1 of 2Next โ†’