954 articles tagged with #llm. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers propose semantic caching solutions for large language models to improve response times and reduce costs by reusing semantically similar requests. The study proves that optimal offline semantic caching is NP-hard and introduces polynomial-time heuristics and online policies combining recency, frequency, and locality factors.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers propose LEAP, a new framework for detecting AI hallucinations using efficient small models that can dynamically adapt verification strategies. The system uses a teacher-student approach where a powerful model trains smaller ones to detect false outputs, addressing a critical barrier to safe AI deployment in production environments.
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง Researchers analyzed 9,705 AI incident reports to create an expanded taxonomy of real-world AI risk mitigation strategies, identifying four new categories of responses including corrective actions, legal enforcement, financial controls, and avoidance tactics. The study expands existing mitigation frameworks by 67% and provides structured guidance for preventing cascading AI system failures in high-stakes deployments.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduce HumanLM, a novel AI training framework that creates user simulators by aligning psychological states rather than just imitating response patterns. The system achieved 16.3% improvement in alignment scores across six datasets with 26k users and 216k responses, demonstrating superior ability to simulate real human behavior.
AIBearisharXiv โ CS AI ยท Mar 56/10
๐ง Research comparing four state-of-the-art language models (GPT-5, Gemini 2.5 Pro, Claude Sonnet 4.5, and Centaur) to humans in goal selection tasks reveals substantial divergence in behavior. While humans explore diverse approaches and learn gradually, the AI models tend to exploit single solutions or show poor performance, raising concerns about using current LLMs as proxies for human decision-making in critical applications.
๐ง Claude๐ง Gemini
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers developed MA-RAG, a Multi-Round Agentic RAG framework that improves medical AI reasoning by iteratively refining responses through conflict detection and external evidence retrieval. The system achieved a substantial +6.8 point accuracy improvement over baseline models across 7 medical Q&A benchmarks by addressing hallucinations and outdated knowledge in healthcare AI applications.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers developed CoCo-TAMP, a robot planning framework that uses large language models to improve state estimation in partially observable environments. The system leverages LLMs' common-sense reasoning to predict object locations and co-locations, achieving 62-73% reduction in planning time compared to baseline methods.
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง Researchers have developed DBench-Bio, a dynamic benchmark system that automatically evaluates AI's ability to discover new biological knowledge using a three-stage pipeline of data acquisition, question-answer extraction, and quality filtering. The benchmark addresses the critical problem of data contamination in static datasets and provides monthly updates across 12 biomedical domains, revealing current limitations in state-of-the-art AI models' knowledge discovery capabilities.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce SHE (Stepwise Hybrid Examination), a new reinforcement learning framework that improves AI-powered e-commerce search relevance prediction. The framework addresses limitations in existing training methods by using step-level rewards and hybrid verification to enhance both accuracy and interpretability of search results.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce TATRA, a training-free prompting method for Large Language Models that creates instance-specific few-shot prompts without requiring labeled training data. The method achieves state-of-the-art performance on mathematical reasoning benchmarks like GSM8K and DeepMath, matching or outperforming existing prompt optimization methods that rely on expensive training processes.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce TTSR, a new framework that enables AI models to improve their reasoning abilities during test time by having a single model alternate between student and teacher roles. The system allows models to learn from their mistakes by analyzing failed reasoning attempts and generating targeted practice questions for continuous improvement.
AIBearisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers have identified 'preference leakage,' a contamination problem in LLM-as-a-judge systems where evaluator models show bias toward related data generator models. The study found this bias occurs when judge and generator LLMs share relationships like being the same model, having inheritance connections, or belonging to the same model family.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers propose a geometric framework showing how large language models 'think' through representation space as flows, with logical statements acting as controllers of these flows' velocities. The study provides evidence that LLMs can internalize logical invariants through next-token prediction training, challenging the 'stochastic parrot' criticism and suggesting universal representational laws underlying machine understanding.
AIBearisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers developed a new AI safety attack method using optimal transport theory that achieves 11% higher success rates in bypassing language model safety mechanisms compared to existing approaches. The study reveals that AI safety refusal mechanisms are localized to specific network layers rather than distributed throughout the model, suggesting current alignment methods may be more vulnerable than previously understood.
๐ข Perplexity๐ง Llama
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduce MACC (Multi-Agent Collaborative Competition), a new institutional architecture that combines multiple AI agents based on large language models to improve scientific discovery. The system addresses limitations of single-agent approaches by incorporating incentive mechanisms, shared workspaces, and institutional design principles to enhance transparency, reproducibility, and exploration efficiency in scientific research.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduce MMAI Gym for Science, a training framework for molecular foundation models in drug discovery. Their Liquid Foundation Model (LFM) outperforms larger general-purpose models on drug discovery tasks while being more efficient and specialized for molecular applications.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce Concentration-Alignment Transforms (CAT), a new method to reduce quantization error in large language and vision models by improving both weight/activation concentration and alignment. The technique consistently matches or outperforms existing quantization methods at 4-bit precision across several LLMs.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers discovered that Large Language Models become increasingly sparse in their internal representations when handling more difficult or out-of-distribution tasks. This sparsity mechanism appears to be an adaptive response that helps stabilize reasoning under challenging conditions, leading to the development of a new learning strategy called Sparsity-Guided Curriculum In-Context Learning (SG-ICL).
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduce Adversarially-Aligned Jacobian Regularization (AAJR), a new method to improve the robustness of autonomous AI agent systems by controlling sensitivity along adversarial directions rather than globally. This approach maintains better performance while ensuring stability in multi-agent AI ecosystems compared to existing methods.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce MASS, a meta-learning framework that enables large language models to self-adapt at test time by generating synthetic training data and performing targeted self-updates. The system uses bilevel optimization to meta-learn data-attribution signals and optimize synthetic data through scalable meta-gradients, showing effectiveness in mathematical reasoning tasks.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง MemSifter is a new AI framework that uses smaller proxy models to handle memory retrieval for large language models, addressing computational costs in long-term memory tasks. The system uses reinforcement learning to optimize retrieval accuracy and has been open-sourced with demonstrated performance improvements on benchmark tests.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduce AgentSelect, a comprehensive benchmark for recommending AI agent configurations based on narrative queries. The benchmark aggregates over 111,000 queries and 107,000 deployable agents from 40+ sources to address the critical gap in selecting optimal LLM agent setups for specific tasks.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers have introduced Mozi, a dual-layer architecture designed to make AI agents more reliable for drug discovery by implementing governance controls and structured workflows. The system addresses critical issues of unconstrained tool use and poor long-term reliability that have limited LLM deployment in pharmaceutical research.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduce GraphMERT, an 80M-parameter AI model that efficiently extracts reliable knowledge graphs from unstructured text data. The system outperforms much larger language models like Qwen3-32B in generating factually accurate and semantically valid knowledge graphs, achieving 69.8% FActScore versus 40.2% for the baseline.
AINeutralarXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce SafeCRS, a safety-aware training framework for LLM-based conversational recommender systems that addresses personalized safety vulnerabilities. The system reduces safety violation rates by up to 96.5% while maintaining recommendation quality by respecting individual user constraints like trauma triggers and phobias.