253 articles tagged with #benchmark. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv – CS AI · Mar 277/10
🧠Researchers introduce ARC-AGI-3, a new benchmark for testing agentic AI systems that focuses on fluid adaptive intelligence without relying on language or external knowledge. While humans can solve 100% of the benchmark's abstract reasoning tasks, current frontier AI systems score below 1% as of March 2026.
AINeutralarXiv – CS AI · Mar 277/10
🧠Researchers introduced WebTestBench, a new benchmark for evaluating automated web testing using AI agents and large language models. The study reveals significant gaps between current AI capabilities and industrial deployment needs, with LLMs struggling with test completeness, defect detection, and long-term interaction reliability.
AIBearishDecrypt · Mar 267/10
🧠A new AI benchmark called ARC-AGI-3 was released the same week Jensen Huang claimed AGI was achieved, showing dramatically poor performance from leading AI models. While humans scored 100% on the benchmark, advanced models like Gemini and GPT scored less than 0.4%, suggesting artificial general intelligence remains far from reality.
🧠 GPT-5🧠 Gemini
AIBearisharXiv – CS AI · Mar 267/10
🧠Researchers introduced EnterpriseArena, the first benchmark testing whether AI agents can function as CFOs by allocating resources in complex enterprise environments over 132 months. Testing on eleven advanced LLMs revealed poor performance, with only 16% of runs surviving the full simulation period, highlighting significant capability gaps in long-term resource allocation under uncertainty.
AIBearishFortune Crypto · Mar 177/10
🧠Legendary venture capitalist Bill Gurley from Benchmark warns that the AI bubble is nearing its end, predicting a market reset when companies 'trip and run out of money.' He believes this critical moment is approaching for the overheated AI sector.
AINeutralarXiv – CS AI · Mar 177/10
🧠Researchers introduced WebCoderBench, the first comprehensive benchmark for evaluating web application generation by large language models, featuring 1,572 real-world user requirements and 24 evaluation metrics. The benchmark tests 12 representative LLMs and shows no single model dominates across all metrics, providing opportunities for targeted improvements.
AIBullisharXiv – CS AI · Mar 177/10
🧠Researchers developed AD-Copilot, a specialized multimodal AI assistant for industrial anomaly detection that outperforms existing models and even human experts. The system uses a novel visual comparison approach and achieved 82.3% accuracy on benchmarks, representing up to 3.35x improvement over baselines.
🏢 Microsoft
AIBullisharXiv – CS AI · Mar 177/10
🧠Researchers introduce SuperLocalMemory V3, a new mathematical framework for AI agent memory systems using information geometry and sheaf theory. The system achieves 87.7% accuracy with cloud augmentation and offers a zero-LLM configuration that complies with EU AI Act data sovereignty requirements.
AINeutralarXiv – CS AI · Mar 177/10
🧠Researchers introduce AVA-Bench, a new benchmark that evaluates vision foundation models (VFMs) by testing 14 distinct atomic visual abilities like localization and depth estimation. This approach provides more precise assessment than traditional VQA benchmarks and reveals that smaller 0.5B language models can evaluate VFMs as effectively as 7B models while using 8x fewer GPU resources.
AIBullisharXiv – CS AI · Mar 177/10
🧠Researchers introduced PriCoder, a new approach that improves Large Language Models' ability to generate code using private library APIs by over 20%. The method uses automatically synthesized training data through graph-based operators to teach LLMs private library usage, addressing a key limitation in current AI coding capabilities.
AIBullisharXiv – CS AI · Mar 177/10
🧠Researchers introduce BevAD, a new lightweight end-to-end autonomous driving architecture that achieves 72.7% success rate on the Bench2Drive benchmark. The study systematically analyzes architectural patterns in closed-loop driving performance, revealing limitations of open-loop dataset approaches and demonstrating strong data-scaling behavior through pure imitation learning.
AINeutralarXiv – CS AI · Mar 177/10
🧠Researchers introduce CCTU, a new benchmark for evaluating large language models' ability to use tools under complex constraints. The study reveals that even state-of-the-art LLMs achieve less than 20% task completion rates when strict constraint adherence is required, with models violating constraints in over 50% of cases.
AIBearisharXiv – CS AI · Mar 167/10
🧠Researchers have released MalURLBench, the first benchmark to evaluate how LLM-based web agents handle malicious URLs, revealing significant vulnerabilities across 12 popular models. The study found that existing AI agents struggle to detect disguised malicious URLs and proposed URLGuard as a defensive solution.
AIBearisharXiv – CS AI · Mar 167/10
🧠Researchers introduced OffTopicEval, a benchmark revealing that all major LLMs suffer from poor operational safety, with even top performers like Qwen-3 and Mistral achieving only 77-80% accuracy in staying on-topic for specific use cases. The study proposes prompt-based steering methods that can improve performance by up to 41%, highlighting critical safety gaps in current AI deployment.
🧠 Llama
AIBearisharXiv – CS AI · Mar 167/10
🧠Researchers introduced CoRE, a benchmark testing whether large language models can reason about human emotions through cognitive dimensions rather than just labels. The study found that while LLMs capture systematic relations between cognitive appraisals and emotions, they show misalignment with human judgments and instability across different contexts.
AIBullisharXiv – CS AI · Mar 167/10
🧠Researchers developed a new reinforcement learning approach for training diffusion language models that uses entropy-guided step selection and stepwise advantages to overcome challenges with sequence-level likelihood calculations. The method achieves state-of-the-art results on coding and logical reasoning benchmarks while being more computationally efficient than existing approaches.
AINeutralarXiv – CS AI · Mar 127/10
🧠Researchers conducted comprehensive benchmarks of LLM inference on AMD Instinct MI325X GPUs, testing models from 235B to 1 trillion parameters. The study reveals that architecture-aware optimization is critical, with different model types requiring specific configurations for optimal performance on AMD hardware.
🧠 Llama
AINeutralarXiv – CS AI · Mar 127/10
🧠Researchers developed DeliberationBench, a new benchmark to assess how large language models influence users' opinions on policy matters. A study of 4,088 participants discussing 65 policy proposals with six frontier LLMs found that these models have substantial influence that appears to align with democratically legitimate deliberative processes.
AINeutralarXiv – CS AI · Mar 117/10
🧠Researchers have developed an open-source benchmark dataset to evaluate AI systems' compliance with the EU AI Act, specifically focusing on NLP and RAG systems. The dataset enables automated assessment of risk classification, article retrieval, and question-answering tasks, achieving 0.87 and 0.85 F1-scores for prohibited and high-risk scenarios.
AINeutralarXiv – CS AI · Mar 117/10
🧠Researchers introduce MiniAppBench, a new benchmark for evaluating Large Language Models' ability to generate interactive HTML applications rather than static text responses. The benchmark includes 500 real-world tasks and an agentic evaluation framework called MiniAppEval that uses browser automation for testing.
AIBullisharXiv – CS AI · Mar 117/10
🧠Researchers introduce SATURN, a new reinforcement learning framework that uses Boolean Satisfiability (SAT) problems to improve large language models' reasoning capabilities. The framework addresses key limitations in existing RL approaches by enabling scalable task construction, automated verification, and precise difficulty control through curriculum learning.
AINeutralarXiv – CS AI · Mar 117/10
🧠Researchers introduce OOD-MMSafe, a new benchmark revealing that current Multimodal Large Language Models fail to identify hidden safety risks up to 67.5% of the time. They developed CASPO framework which dramatically reduces failure rates to under 8% for risk identification in consequence-driven safety scenarios.
AINeutralarXiv – CS AI · Mar 97/10
🧠Researchers introduced LLMTM, a comprehensive benchmark to evaluate Large Language Models' performance on temporal motif analysis in dynamic graphs. The study tested nine different LLMs and developed a structure-aware dispatcher that balances accuracy with cost-effectiveness for graph analysis tasks.
🧠 GPT-4
AIBullisharXiv – CS AI · Mar 67/10
🧠Researchers introduce the Dynamic Behavioral Constraint (DBC) benchmark, a new governance framework for large language models that reduces AI risk exposure by 36.8% through structured behavioral controls applied at inference time. The system achieves high EU AI Act compliance scores and represents a model-agnostic approach to AI safety that can be audited and mapped to different jurisdictions.
AIBullisharXiv – CS AI · Mar 57/10
🧠Researchers have released RoboCasa365, a large-scale simulation benchmark featuring 365 household tasks across 2,500 kitchen environments with over 600 hours of human demonstration data. The platform is designed to train and evaluate generalist robots for everyday tasks, providing insights into factors affecting robot performance and generalization capabilities.