y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#evaluation-framework News & Analysis

20 articles tagged with #evaluation-framework. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

20 articles
AINeutralarXiv โ€“ CS AI ยท 2d ago7/10
๐Ÿง 

The Long-Horizon Task Mirage? Diagnosing Where and Why Agentic Systems Break

Researchers introduce HORIZON, a diagnostic benchmark for identifying and analyzing why large language model agents fail at long-horizon tasks requiring extended action sequences. By evaluating state-of-the-art models across multiple domains and proposing an LLM-as-a-Judge attribution pipeline, the study provides systematic methodology for understanding agent limitations and improving reliability.

๐Ÿง  GPT-5๐Ÿง  Claude
AI ร— CryptoNeutralarXiv โ€“ CS AI ยท Apr 107/10
๐Ÿค–

Blockchain and AI: Securing Intelligent Networks for the Future

A comprehensive academic synthesis examines how blockchain and AI technologies can be integrated to secure intelligent networks across IoT, critical infrastructure, and healthcare. The paper introduces a taxonomy, integration patterns, and the BASE evaluation blueprint to standardize security assessments, revealing that while the conceptual alignment is strong, real-world implementations remain largely prototype-stage.

AIBearisharXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

Brittlebench: Quantifying LLM robustness via prompt sensitivity

Researchers introduce Brittlebench, a new evaluation framework that reveals frontier AI models experience up to 12% performance degradation when faced with minor prompt variations like typos or rephrasing. The study shows that semantics-preserving input perturbations can account for up to half of a model's performance variance, highlighting significant robustness issues in current language models.

AINeutralarXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

Eva-VLA: Evaluating Vision-Language-Action Models' Robustness Under Real-World Physical Variations

Researchers introduced Eva-VLA, the first unified framework to systematically evaluate the robustness of Vision-Language-Action models for robotic manipulation under real-world physical variations. Testing revealed OpenVLA exhibits over 90% failure rates across three physical variations, exposing critical weaknesses in current VLA models when deployed outside laboratory conditions.

AINeutralarXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Towards Personalized Deep Research: Benchmarks and Evaluations

Researchers introduce PDR-Bench, the first benchmark for evaluating personalization in Deep Research Agents (DRAs), featuring 250 realistic user-task queries across 10 domains. The benchmark uses a new PQR Evaluation Framework to measure personalization alignment, content quality, and factual reliability in AI research assistants.

AIBullisharXiv โ€“ CS AI ยท Feb 277/107
๐Ÿง 

General Agent Evaluation

Researchers have developed Exgentic, a new framework for evaluating general-purpose AI agents that can perform tasks across different environments without domain-specific tuning. The study benchmarked five prominent agent implementations and found that general agents can achieve performance comparable to specialized agents, establishing the first Open General Agent Leaderboard.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

COMPOSITE-Stem

Researchers introduced COMPOSITE-STEM, a new benchmark containing 70 expert-written scientific tasks across physics, biology, chemistry, and mathematics to evaluate AI agents. The top-performing model achieved only 21% accuracy, indicating the benchmark effectively measures capabilities beyond current AI reach and addresses the saturation of existing evaluation frameworks.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

NovBench: Evaluating Large Language Models on Academic Paper Novelty Assessment

Researchers introduced NovBench, the first large-scale benchmark for evaluating how well large language models can assess research novelty in academic papers. The benchmark comprises 1,684 paper-review pairs from a leading NLP conference and reveals that current LLMs struggle with scientific novelty comprehension despite promise in peer review support.

AINeutralarXiv โ€“ CS AI ยท Apr 106/10
๐Ÿง 

Large Language Models for Outpatient Referral: Problem Definition, Benchmarking and Challenges

Researchers have developed a comprehensive evaluation framework for Large Language Models applied to outpatient referral systems in healthcare, revealing that LLMs offer limited advantages over simpler BERT-like models in static referral tasks but demonstrate potential in interactive dialogue scenarios. The study addresses the absence of standardized evaluation criteria for assessing LLM effectiveness in dynamic healthcare settings.

AINeutralarXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Evaluation of Audio Language Models for Fairness, Safety, and Security

Researchers introduce a structural taxonomy and unified evaluation framework for Audio Large Language Models (ALLMs) to assess fairness, safety, and security. The study reveals systematic differences in how ALLMs handle audio versus text inputs, with FSS behavior closely tied to acoustic information integration methods.

AINeutralarXiv โ€“ CS AI ยท Mar 166/10
๐Ÿง 

Budget-Sensitive Discovery Scoring: A Formally Verified Framework for Evaluating AI-Guided Scientific Selection

Researchers introduce Budget-Sensitive Discovery Score (BSDS), a formally verified framework for evaluating AI-guided scientific candidate selection under budget constraints. Testing on drug discovery datasets reveals that simple random forest models outperform large language models, with LLMs providing no marginal value over existing trained classifiers.

AINeutralarXiv โ€“ CS AI ยท Mar 37/109
๐Ÿง 

Property-Driven Evaluation of GNN Expressiveness at Scale: Datasets, Framework, and Study

Researchers developed a comprehensive evaluation framework for Graph Neural Networks (GNNs) using formal specification methods, creating 336 new datasets to test GNN expressiveness across 16 fundamental graph properties. The study reveals that no single pooling approach consistently performs well across all properties, with attention-based pooling excelling in generalization while second-order pooling provides better sensitivity.

AIBullisharXiv โ€“ CS AI ยท Mar 37/106
๐Ÿง 

MOSAIC: A Unified Platform for Cross-Paradigm Comparison and Evaluation of Homogeneous and Heterogeneous Multi-Agent RL, LLM, VLM, and Human Decision-Makers

MOSAIC is a new open-source platform that enables cross-paradigm comparison and evaluation of different AI agents including reinforcement learning, large language models, vision-language models, and human decision-makers within the same environment. The platform introduces three key technical contributions: an IPC-based worker protocol, operator abstraction for unified interfaces, and a deterministic evaluation framework for reproducible research.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

ScholarEval: Research Idea Evaluation Grounded in Literature

Researchers introduce ScholarEval, a retrieval-augmented framework for evaluating AI-generated research ideas based on soundness and contribution metrics. The system outperformed OpenAI's o1-mini-deep-research baseline across multiple evaluation criteria in testing with 117 expert-annotated research ideas across four scientific disciplines.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

A High-Quality Dataset and Reliable Evaluation for Interleaved Image-Text Generation

Researchers introduced InterSyn, a 1.8M sample dataset designed to improve Large Multimodal Models' ability to generate interleaved image-text content. The dataset includes a new evaluation framework called SynJudge that measures four key performance metrics, with experiments showing significant improvements even with smaller 25K-50K sample subsets.

AINeutralHugging Face Blog ยท Apr 166/108
๐Ÿง 

Introducing HELMET: Holistically Evaluating Long-context Language Models

HELMET is a new holistic evaluation framework for assessing long-context language models across multiple dimensions and use cases. The framework aims to provide comprehensive benchmarking capabilities for AI models that can process extended text sequences.

AINeutralarXiv โ€“ CS AI ยท Mar 264/10
๐Ÿง 

No Single Metric Tells the Whole Story: A Multi-Dimensional Evaluation Framework for Uncertainty Attributions

Researchers propose a new framework for evaluating uncertainty attribution methods in explainable AI, addressing inconsistent evaluation practices in the field. The study introduces five key properties including a new 'conveyance' metric and demonstrates that gradient-based methods outperform perturbation-based approaches across multiple evaluation criteria.