12,960 AI articles curated from 50+ sources with AI-powered sentiment analysis, importance scoring, and key takeaways.
AINeutralarXiv – CS AI · Mar 37/108
🧠Researchers introduce PhotoBench, the first benchmark for personalized photo retrieval using authentic personal albums rather than web images. The study reveals critical limitations in current AI systems, including modality gaps in unified embedding models and poor tool orchestration in agentic systems.
AINeutralarXiv – CS AI · Mar 37/106
🧠Researchers have created the Sentience Readiness Index (SRI) to measure how prepared 31 countries are for the possibility of AI achieving consciousness. No nation scored above 'Partially Prepared,' with the UK leading at 49/100, revealing significant gaps in institutional, professional, and cultural infrastructure needed to handle potentially sentient AI systems.
AINeutralarXiv – CS AI · Mar 36/105
🧠Researchers introduced Spoof-SUPERB, a new benchmark for evaluating self-supervised learning models' ability to detect audio deepfakes. The study tested 20 SSL models and found that large-scale discriminative models like XLS-R and WavLM Large consistently outperformed others, especially under acoustic degradations.
AIBullisharXiv – CS AI · Mar 36/108
🧠Researchers propose ATA, a training-free framework that improves Vision-Language-Action (VLA) models through implicit reasoning without requiring additional data or annotations. The approach uses attention-guided and action-guided strategies to enhance visual inputs, achieving better task performance while maintaining inference efficiency.
AIBullisharXiv – CS AI · Mar 36/106
🧠Researchers introduce 3R, a new RAG-based framework that optimizes prompts for text-to-video generation models without requiring model retraining. The system uses three key strategies to improve video quality: RAG-based modifier extraction, diffusion-based preference optimization, and temporal frame interpolation for better consistency.
AIBullisharXiv – CS AI · Mar 37/109
🧠Researchers have developed MM-Mem, a new pyramidal multimodal memory architecture that enables AI systems to better understand long-horizon videos by mimicking human cognitive memory processes. The system addresses current limitations in multimodal large language models by creating a hierarchical memory structure that progressively distills detailed visual information into high-level semantic understanding.
AIBearisharXiv – CS AI · Mar 37/108
🧠Researchers have discovered VidDoS, a new universal attack framework that can severely degrade Video-based Large Language Models by causing extreme computational resource exhaustion. The attack increases token generation by over 205x and inference latency by more than 15x, creating critical safety risks in real-world applications like autonomous driving.
AIBullisharXiv – CS AI · Mar 36/107
🧠Researchers developed a Mean-Flow based One-Step Vision-Language-Action (VLA) approach that dramatically improves robotic manipulation efficiency by eliminating iterative sampling requirements. The new method achieves 8.7x faster generation than SmolVLA and 83.9x faster than Diffusion Policy in real-world robotic experiments.
AIBullisharXiv – CS AI · Mar 36/107
🧠Researchers have developed RGLM, a new approach to improve how large language models understand and process graph data by incorporating explicit graph supervision alongside text instructions. The method addresses limitations in existing Graph-Tokenizing LLMs that rely too heavily on text supervision, leading to underutilization of graph context.
AIBullisharXiv – CS AI · Mar 37/107
🧠Researchers developed a method for creating synthetic instruction datasets to improve domain-specific LLMs, demonstrating with a 9.5 billion token Japanese financial dataset. The approach enhances both domain expertise and reasoning capabilities, with models and datasets being open-sourced for broader use.
AIBearisharXiv – CS AI · Mar 36/107
🧠Researchers created PanCanBench, a comprehensive benchmark evaluating 22 large language models on pancreatic cancer-related patient questions, revealing significant variations in clinical accuracy and high hallucination rates. The study found that even top-performing models like GPT-4o and Gemini-2.5 Pro had hallucination rates of 6%, while newer reasoning-optimized models didn't consistently improve factual accuracy.
AINeutralarXiv – CS AI · Mar 37/108
🧠Researchers propose a new method called total Variation-based Advantage aligned Constrained policy Optimization to address policy lag issues in distributed reinforcement learning systems. The approach aims to improve performance when scaling on-policy learning algorithms by mitigating the mismatch between behavior and learning policies during high-frequency updates.
AIBullisharXiv – CS AI · Mar 36/109
🧠Researchers introduce In-Context Policy Optimization (ICPO), a new method that allows AI models to improve their responses during inference through multi-round self-reflection without parameter updates. The practical ME-ICPO algorithm demonstrates competitive performance on mathematical reasoning tasks while maintaining affordable inference costs.
AIBullisharXiv – CS AI · Mar 37/107
🧠Researchers introduce Pri4R, a new approach that enhances Vision-Language-Action (VLA) models by incorporating 4D spatiotemporal understanding during training. The method adds a lightweight point track head that predicts 3D trajectories, improving physical world understanding while maintaining the original architecture during inference with no computational overhead.
AIBearisharXiv – CS AI · Mar 36/108
🧠Research reveals that Large Language Model (LLM) self-explanations fail semantic invariance testing, showing that AI models' self-reports change based on how tasks are framed rather than actual task performance. Four frontier AI models demonstrated unreliable self-reporting when faced with semantically different but functionally identical tool descriptions, raising questions about using model self-reports as evidence of capability.
AIBullisharXiv – CS AI · Mar 36/105
🧠Researchers introduce 'semi-formal reasoning' for LLM agents to analyze code semantics without execution, showing significant accuracy improvements across multiple tasks. The methodology achieves 88-93% accuracy on patch verification and 87% on code question answering, potentially enabling practical applications in automated code review and static analysis.
AIBullisharXiv – CS AI · Mar 37/107
🧠Researchers propose Causal Neural Probabilistic Circuits (CNPC), a new AI model that enhances interpretable machine learning by incorporating causal dependencies between concepts. The model allows domain experts to make corrections that properly propagate through causal relationships, achieving higher accuracy than baseline models across benchmark datasets.
AIBullisharXiv – CS AI · Mar 37/106
🧠MOSAIC is a new open-source platform that enables cross-paradigm comparison and evaluation of different AI agents including reinforcement learning, large language models, vision-language models, and human decision-makers within the same environment. The platform introduces three key technical contributions: an IPC-based worker protocol, operator abstraction for unified interfaces, and a deterministic evaluation framework for reproducible research.
AIBullisharXiv – CS AI · Mar 36/106
🧠Researchers introduce GlassMol, a new interpretable AI model for molecular property prediction that addresses the black-box problem in drug discovery. The model uses Concept Bottleneck Models with automated concept curation and LLM-guided selection, achieving performance that matches or exceeds traditional black-box models across thirteen benchmarks.
AIBullisharXiv – CS AI · Mar 37/106
🧠Researchers introduce SEKA and AdaSEKA, new training-free methods for attention steering in AI models that work with memory-efficient implementations like FlashAttention. These techniques enable better prompt highlighting by directly editing key embeddings using spectral decomposition, offering significant performance improvements with lower computational overhead.
AIBearisharXiv – CS AI · Mar 37/109
🧠A study reveals that safety-aligned large language models exhibit "Defensive Refusal Bias," refusing legitimate cybersecurity defense tasks 2.72x more often when they contain security-sensitive keywords. The research found particularly high refusal rates for critical defensive operations like system hardening (43.8%) and malware analysis (34.3%), suggesting current AI safety measures rely on semantic similarity rather than understanding intent.
AINeutralarXiv – CS AI · Mar 37/108
🧠The MAMA-MIA Challenge introduced a large-scale benchmark for AI-powered breast cancer tumor segmentation and treatment response prediction using MRI data from 1,506 US patients for training and 574 European patients for testing. Results from 26 international teams revealed significant performance variability and trade-offs between accuracy and fairness across demographic subgroups when AI models were tested across different institutions and continents.
AIBullisharXiv – CS AI · Mar 36/106
🧠Researchers developed KG-Followup, a knowledge graph-augmented large language model system that generates medical follow-up questions for pre-diagnostic assessment. The system combines structured medical domain knowledge with LLMs to improve clinical diagnosis efficiency, outperforming existing methods by 5-8% in recall benchmarks.
AINeutralarXiv – CS AI · Mar 36/106
🧠Researchers identified Self-Anchoring Calibration Drift (SACD), where large language models show systematic confidence changes when building on their own outputs in multi-turn conversations. Testing Claude Sonnet 4.6, Gemini 3.1 Pro, and GPT-5.2 revealed model-specific patterns, with Claude showing decreasing confidence and significant calibration errors, while GPT-5.2 exhibited opposite behavior in open-ended domains.
$NEAR
AIBullisharXiv – CS AI · Mar 36/106
🧠Researchers developed TARSE, a new AI system for clinical decision-making that retrieves relevant medical skills and experiences from curated libraries to improve reasoning accuracy. The system performs test-time adaptation to align language models with clinically valid logic, showing improvements over existing medical AI baselines in question-answering benchmarks.