y0news
#ai-evaluation6 articles
6 articles
AINeutralarXiv โ€“ CS AI ยท 6h ago5
๐Ÿง 

CIRCLE: A Framework for Evaluating AI from a Real-World Lens

Researchers propose CIRCLE, a six-stage framework for evaluating AI systems through real-world deployment outcomes rather than abstract model performance metrics. The framework aims to bridge the gap between theoretical AI capabilities and actual materialized effects by providing systematic evidence for decision-makers outside the AI development stack.

AIBearisharXiv โ€“ CS AI ยท 6h ago4
๐Ÿง 

Humans and LLMs Diverge on Probabilistic Inferences

Researchers created ProbCOPA, a dataset testing probabilistic reasoning in humans versus AI models, finding that state-of-the-art LLMs consistently fail to match human judgment patterns. The study reveals fundamental differences in how humans and AI systems process non-deterministic inferences, highlighting limitations in current AI reasoning capabilities.

AIBullisharXiv โ€“ CS AI ยท 6h ago6
๐Ÿง 

LLM-Driven Multi-Turn Task-Oriented Dialogue Synthesis for Realistic Reasoning

Researchers propose an LLM-driven framework for generating multi-turn task-oriented dialogues to create more realistic reasoning benchmarks. The framework addresses limitations in current AI evaluation methods by producing synthetic datasets that better reflect real-world complexity and contextual coherence.

AINeutralarXiv โ€“ CS AI ยท 6h ago3
๐Ÿง 

DLEBench: Evaluating Small-scale Object Editing Ability for Instruction-based Image Editing Model

Researchers introduce DLEBench, the first benchmark specifically designed to evaluate instruction-based image editing models' ability to edit small-scale objects that occupy only 1%-10% of image area. Testing on 10 models revealed significant performance gaps in small object editing, highlighting a critical limitation in current AI image editing capabilities.

AIBullisharXiv โ€“ CS AI ยท 6h ago3
๐Ÿง 

Resources for Automated Evaluation of Assistive RAG Systems that Help Readers with News Trustworthiness Assessment

Researchers developed the TREC 2025 DRAGUN Track to evaluate AI systems that help readers assess news trustworthiness through automated report generation. The initiative created reusable evaluation resources including human-assessed rubrics and an AutoJudge system that correlates well with human evaluations for RAG-based news analysis tools.

AIBullisharXiv โ€“ CS AI ยท 6h ago7
๐Ÿง 

Capabilities Ain't All You Need: Measuring Propensities in AI

Researchers introduce the first formal framework for measuring AI propensities - the tendencies of models to exhibit particular behaviors - going beyond traditional capability measurements. The new bilogistic approach successfully predicts AI behavior on held-out tasks and shows stronger predictive power when combining propensities with capabilities than using either measure alone.