y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#retrieval News & Analysis

13 articles tagged with #retrieval. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

13 articles
AIBullisharXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

MemMachine: A Ground-Truth-Preserving Memory System for Personalized AI Agents

MemMachine is an open-source memory system for AI agents that preserves conversational ground truth and achieves superior accuracy-efficiency tradeoffs compared to existing solutions. The system integrates short-term, long-term episodic, and profile memory while using 80% fewer input tokens than comparable systems like Mem0.

๐Ÿง  GPT-4๐Ÿง  GPT-5
AINeutralarXiv โ€“ CS AI ยท Mar 46/104
๐Ÿง 

Diagnosing Retrieval vs. Utilization Bottlenecks in LLM Agent Memory

Researchers analyzed memory systems in LLM agents and found that retrieval methods are more critical than write strategies for performance. Simple raw chunk storage matched expensive alternatives, suggesting current memory pipelines may discard useful context that retrieval systems cannot compensate for.

AIBullisharXiv โ€“ CS AI ยท Feb 277/108
๐Ÿง 

RAGdb: A Zero-Dependency, Embeddable Architecture for Multimodal Retrieval-Augmented Generation on the Edge

Researchers introduce RAGdb, a revolutionary architecture that consolidates Retrieval-Augmented Generation into a single SQLite container, eliminating the need for cloud infrastructure and GPUs. The system achieves 100% entity retrieval accuracy while reducing disk footprint by 99.5% compared to traditional Docker-based RAG stacks, enabling truly portable AI applications for edge computing and privacy-sensitive environments.

AIBullisharXiv โ€“ CS AI ยท Apr 76/10
๐Ÿง 

GROUNDEDKG-RAG: Grounded Knowledge Graph Index for Long-document Question Answering

Researchers introduced GroundedKG-RAG, a new retrieval-augmented generation system that creates knowledge graphs directly grounded in source documents to improve long-document question answering. The system reduces resource consumption and hallucinations while maintaining accuracy comparable to state-of-the-art models at lower cost.

AIBullisharXiv โ€“ CS AI ยท Mar 266/10
๐Ÿง 

MDKeyChunker: Single-Call LLM Enrichment with Rolling Keys and Key-Based Restructuring for High-Accuracy RAG

Researchers introduce MDKeyChunker, a three-stage pipeline that improves RAG (Retrieval-Augmented Generation) systems by using structure-aware chunking of Markdown documents, single-call LLM enrichment, and semantic key-based restructuring. The system achieves superior retrieval performance with Recall@5=1.000 using BM25 over structural chunks, significantly improving upon traditional fixed-size chunking methods.

๐Ÿข OpenAI
AIBullisharXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

GlobalRAG: Enhancing Global Reasoning in Multi-hop Question Answering via Reinforcement Learning

GlobalRAG is a new reinforcement learning framework that significantly improves multi-hop question answering by decomposing questions into subgoals and coordinating retrieval with reasoning. The system achieves 14.2% average improvements in performance metrics while using only 42% of the training data required by baseline models.

AIBullisharXiv โ€“ CS AI ยท Mar 36/106
๐Ÿง 

TARSE: Test-Time Adaptation via Retrieval of Skills and Experience for Reasoning Agents

Researchers developed TARSE, a new AI system for clinical decision-making that retrieves relevant medical skills and experiences from curated libraries to improve reasoning accuracy. The system performs test-time adaptation to align language models with clinically valid logic, showing improvements over existing medical AI baselines in question-answering benchmarks.

AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

LLaVE: Large Language and Vision Embedding Models with Hardness-Weighted Contrastive Learning

Researchers introduce LLaVE, a new multimodal embedding model that uses hardness-weighted contrastive learning to better distinguish between positive and negative pairs in image-text tasks. The model achieves state-of-the-art performance on the MMEB benchmark, with LLaVE-2B outperforming previous 7B models and demonstrating strong zero-shot transfer capabilities to video retrieval tasks.

AIBullishHugging Face Blog ยท Mar 226/109
๐Ÿง 

Binary and Scalar Embedding Quantization for Significantly Faster & Cheaper Retrieval

The article discusses binary and scalar embedding quantization techniques that can significantly reduce computational costs and increase speed for retrieval systems. These methods compress high-dimensional vector embeddings while maintaining retrieval performance, making AI search and recommendation systems more efficient and cost-effective.