61 articles tagged with #retrieval-augmented-generation. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv β CS AI Β· 2d ago7/10
π§ Researchers introduce GRIP, a unified framework that integrates retrieval decisions directly into language model generation through control tokens, eliminating the need for external retrieval controllers. The system enables models to autonomously decide when to retrieve information, reformulate queries, and terminate retrieval within a single autoregressive process, achieving competitive performance with GPT-4o while using substantially fewer parameters.
π§ GPT-4
AIBullisharXiv β CS AI Β· 2d ago7/10
π§ Researchers introduce AtlasKV, a parametric knowledge integration method that enables large language models to leverage billion-scale knowledge graphs while consuming less than 20GB of VRAM. Unlike traditional retrieval-augmented generation (RAG) approaches, AtlasKV integrates knowledge directly into LLM parameters without requiring external retrievers or extended context windows, reducing inference latency and computational overhead.
AIBullisharXiv β CS AI Β· 2d ago7/10
π§ Researchers introduce Disco-RAG, a discourse-aware framework that enhances Retrieval-Augmented Generation (RAG) systems by explicitly modeling discourse structures and rhetorical relationships between retrieved passages. The method achieves state-of-the-art results on question answering and summarization tasks without fine-tuning, demonstrating that structural understanding of text significantly improves LLM performance on knowledge-intensive tasks.
AINeutralarXiv β CS AI Β· Apr 67/10
π§ Researchers published a comprehensive technical survey on Large Language Model augmentation strategies, examining methods from in-context learning to advanced Retrieval-Augmented Generation techniques. The study provides a unified framework for understanding how structured context at inference time can overcome LLMs' limitations of static knowledge and finite context windows.
AIBullisharXiv β CS AI Β· Mar 277/10
π§ Researchers introduce WriteBack-RAG, a framework that treats knowledge bases in retrieval-augmented generation systems as trainable components rather than static databases. The method distills relevant information from documents into compact knowledge units, improving RAG performance across multiple benchmarks by an average of +2.14%.
AIBearisharXiv β CS AI Β· Mar 277/10
π§ Researchers have identified a new attack vector called Epistemic Bias Injection (EBI) that manipulates AI language models by injecting factually correct but biased content into retrieval-augmented generation databases. The attack steers model outputs toward specific viewpoints while evading traditional detection methods, though a new defense mechanism called BiasDef shows promise in mitigating these threats.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ Researchers propose pΒ²RAG, a new privacy-preserving Retrieval-Augmented Generation system that supports arbitrary top-k retrieval while being 3-300x faster than existing solutions. The system uses an interactive bisection method instead of sorting and employs secret sharing across two servers to protect user prompts and database content.
$RAG
AIBullisharXiv β CS AI Β· Mar 177/10
π§ Researchers introduce APEX-Searcher, a new framework that enhances large language models' search capabilities through a two-stage approach combining reinforcement learning for strategic planning and supervised fine-tuning for execution. The system addresses limitations in multi-hop question answering by decoupling retrieval processes into planning and execution phases, showing significant improvements across multiple benchmarks.
AINeutralarXiv β CS AI Β· Mar 97/10
π§ Researchers evaluated 34 large language models on radiology questions, finding that agentic retrieval-augmented reasoning systems improve consensus and reliability across different AI models. The study shows these systems reduce decision variability between models and increase robust correctness, though 72% of incorrect outputs still carried moderate to high clinical severity.
AIBullisharXiv β CS AI Β· Mar 97/10
π§ Researchers introduce RAG-Driver, a retrieval-augmented multi-modal large language model designed for autonomous driving that can provide explainable decisions and control predictions. The system addresses data scarcity and generalization challenges in AI-driven autonomous vehicles by using in-context learning and expert demonstration retrieval.
AINeutralarXiv β CS AI Β· Mar 57/10
π§ Researchers propose RAG-X, a diagnostic framework for evaluating retrieval-augmented generation systems in medical AI applications. The study reveals an 'Accuracy Fallacy' showing a 14% gap between perceived system success and actual evidence-based grounding in medical question-answering systems.
AIBullisharXiv β CS AI Β· Mar 56/10
π§ Researchers introduce OSCAR, a new query-dependent online soft compression method for Retrieval-Augmented Generation (RAG) systems that reduces computational overhead while maintaining performance. The method achieves 2-5x speed improvements in inference with minimal accuracy loss across LLMs from 1B to 24B parameters.
π’ Hugging Face
AIBullisharXiv β CS AI Β· Mar 56/10
π§ Researchers demonstrate that coreference resolution significantly improves Retrieval-Augmented Generation (RAG) systems by reducing ambiguity in document retrieval and enhancing question-answering performance. The study finds that smaller language models benefit more from disambiguation processes, with mean pooling strategies showing superior context capturing after coreference resolution.
AIBullisharXiv β CS AI Β· Mar 47/103
π§ Researchers developed a type-aware retrieval-augmented generation (RAG) method that translates natural language requirements into solver-executable optimization code for industrial applications. The method uses a typed knowledge base and dependency closure to ensure code executability, successfully validated on battery production optimization and job scheduling tasks where conventional RAG approaches failed.
AINeutralarXiv β CS AI Β· 1d ago6/10
π§ Researchers propose Opinion-Aware Retrieval-Augmented Generation (RAG) to address a critical bias in current LLM systems that treat subjective content as noise rather than valuable information. By formalizing the distinction between factual queries (epistemic uncertainty) and opinion queries (aleatoric uncertainty), the team develops an architecture that preserves diverse perspectives in knowledge retrieval, demonstrating 26.8% improved sentiment diversity and 42.7% better entity matching on real-world e-commerce data.
AIBullisharXiv β CS AI Β· 2d ago6/10
π§ Researchers introduce MΒ³KG-RAG, a novel multimodal retrieval-augmented generation system that enhances large language models by integrating multi-hop knowledge graphs with audio-visual data. The approach improves reasoning depth and answer accuracy by filtering irrelevant information through a new grounding and pruning mechanism called GRASP.
$KG
AIBullisharXiv β CS AI Β· 2d ago6/10
π§ Researchers introduce CARO, a two-stage training framework that enhances large language models' ability to perform robust content moderation through analogical reasoning. By combining retrieval-augmented generation with direct preference optimization, CARO achieves 24.9% F1 score improvement over state-of-the-art models including DeepSeek R1 and LLaMA Guard on ambiguous moderation cases.
AINeutralarXiv β CS AI Β· 2d ago6/10
π§ Researchers introduce ConflictQA, a benchmark revealing that large language models struggle with conflicting information across different knowledge sources (text vs. knowledge graphs) in retrieval-augmented generation systems. The study proposes XoT, an explanation-based framework to improve faithful reasoning when LLMs encounter contradictory evidence.
AINeutralarXiv β CS AI Β· 2d ago6/10
π§ RAGen is a new framework for generating domain-specific training data to improve Retrieval-Augmented Generation (RAG) systems. The system creates question-answer-context triples using semantic chunking, concept extraction, and Bloom's Taxonomy principles, enabling faster adaptation of LLMs to specialized domains like scientific research and enterprise knowledge bases.
AINeutralarXiv β CS AI Β· 2d ago6/10
π§ Researchers introduce CodaRAG, a framework that enhances Retrieval-Augmented Generation by treating evidence retrieval as active associative discovery rather than passive lookup. The system achieves 7-10% gains in retrieval recall and 3-11% improvements in generation accuracy by consolidating fragmented knowledge, navigating multi-dimensional pathways, and eliminating noise.
AIBullisharXiv β CS AI Β· 3d ago6/10
π§ Researchers introduce RecaLLM, a post-trained language model that addresses the 'lost-in-thought' phenomenon where retrieval performance degrades during extended reasoning chains. The model interleaves explicit in-context retrieval with reasoning steps and achieves strong performance on long-context benchmarks using training data significantly shorter than existing approaches.
AIBullisharXiv β CS AI Β· 3d ago6/10
π§ Researchers introduce VISOR, a new agentic visual retrieval-augmented generation system that improves how AI models reason over multi-page visual documents. By addressing key technical challenges in evidence gathering and context management, VISOR achieves state-of-the-art results on complex visual reasoning tasks.
AINeutralarXiv β CS AI Β· 6d ago6/10
π§ Researchers introduce chain-of-illocution (CoI) prompting to improve source faithfulness in retrieval-augmented language models, achieving up to 63% gains in source adherence for programming education tasks. The study reveals that standard RAG systems exhibit low fidelity to source materials, with non-RAG models performing worse, while a user study confirms improved faithfulness does not compromise user satisfaction.
AINeutralarXiv β CS AI Β· 6d ago6/10
π§ Researchers propose G-Defense, a graph-enhanced framework that uses large language models and retrieval-augmented generation to detect fake news while providing explainable, fine-grained reasoning. The system decomposes news claims into sub-claims, retrieves competing evidence, and generates transparent explanations without requiring verified fact-checking databases.
AINeutralarXiv β CS AI Β· 6d ago6/10
π§ Researchers have developed Luwen, an open-source Chinese legal language model built on Baichuan that uses continual pre-training, supervised fine-tuning, and retrieval-augmented generation to excel at legal tasks. The model outperforms baselines on five legal benchmarks including judgment prediction, judicial examination, and legal reasoning, demonstrating effective domain adaptation for specialized legal applications.