11 articles tagged with #evaluation-metrics. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBearisharXiv โ CS AI ยท Mar 167/10
๐ง Research reveals that AI agents using tools for financial advice can recommend unsafe products while maintaining good quality metrics when tool data is corrupted. The study found that 65-93% of recommendations contained risk-inappropriate products across seven LLMs, yet standard evaluation metrics failed to detect these safety issues.
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง Researchers propose SemKey, a novel framework that addresses key limitations in EEG-to-text decoding by preventing hallucinations and improving semantic fidelity through decoupled guidance objectives. The system redesigns neural encoder-LLM interaction and introduces new evaluation metrics beyond BLEU scores to achieve state-of-the-art performance in brain-computer interfaces.
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduced InEdit-Bench, the first evaluation benchmark specifically designed to test image editing models' ability to reason through intermediate logical pathways in multi-step visual transformations. Testing 14 representative models revealed significant shortcomings in handling complex scenarios requiring dynamic reasoning and procedural understanding.
AIBullisharXiv โ CS AI ยท 4d ago6/10
๐ง Researchers propose Interactive ASR, a new framework that combines semantic-aware evaluation using LLM-as-a-Judge with multi-turn interactive correction to improve automatic speech recognition beyond traditional word error rate metrics. The approach simulates human-like interaction, enabling iterative refinement of recognition outputs across English, Chinese, and code-switching datasets.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers introduce a formal planning framework that maps LLM-based web agents to traditional search algorithms, enabling better diagnosis of failures in autonomous web tasks. The study compares different agent architectures using novel evaluation metrics and a dataset of 794 human-labeled trajectories from WebArena benchmark.
AINeutralarXiv โ CS AI ยท Mar 66/10
๐ง Researchers introduce ICR (Inductive Conceptual Rating), a new qualitative metric for evaluating meaning in large language model text summaries that goes beyond simple word similarity. The study found that while LLMs achieve high linguistic similarity to human outputs, they significantly underperform in semantic accuracy and capturing contextual meanings.
AINeutralarXiv โ CS AI ยท Mar 36/107
๐ง Researchers introduce MC-Search, the first benchmark for evaluating agentic multimodal retrieval-augmented generation (MM-RAG) systems with long, structured reasoning chains. The benchmark reveals systematic issues in current multimodal large language models and introduces Search-Align, a training framework that improves planning and retrieval accuracy.
AINeutralarXiv โ CS AI ยท Mar 36/107
๐ง Researchers developed an event-based evaluation framework for LLM-generated clinical summaries of remote monitoring data, revealing that models with high semantic similarity often fail to capture clinically significant events. A vision-based approach using time-series visualizations achieved the best clinical event alignment with 45.7% abnormality recall.
$NEAR
AINeutralarXiv โ CS AI ยท Feb 276/107
๐ง Researchers have developed SPM-Bench, a PhD-level benchmark for testing large language models on scanning probe microscopy tasks. The benchmark uses automated data synthesis from scientific papers and introduces new evaluation metrics to assess AI reasoning capabilities in specialized scientific domains.
AINeutralarXiv โ CS AI ยท Mar 174/10
๐ง Researchers introduce NV-Bench, the first standardized benchmark for evaluating nonverbal vocalizations in text-to-speech systems. The benchmark includes 1,651 multilingual utterances across 14 categories and proposes new evaluation metrics that show strong correlation with human perception.
AINeutralarXiv โ CS AI ยท Mar 94/10
๐ง Researchers developed a methodology to fine-tune large language models (LLMs) for generating code-switched text between English and Spanish by back-translating natural code-switched sentences into monolingual English. The study found that fine-tuning significantly improves LLMs' ability to generate fluent code-switched text, and that LLM-based evaluation methods align better with human preferences than traditional metrics.