y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#perception News & Analysis

11 articles tagged with #perception. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

11 articles
AIBullisharXiv – CS AI · Apr 77/10
🧠

V-Reflection: Transforming MLLMs from Passive Observers to Active Interrogators

Researchers introduce V-Reflection, a new framework that transforms Multimodal Large Language Models (MLLMs) from passive observers to active interrogators through a 'think-then-look' mechanism. The approach addresses perception-related hallucinations in fine-grained tasks by allowing models to dynamically re-examine visual details during reasoning, showing significant improvements across six perception-intensive benchmarks.

AINeutralarXiv – CS AI · Mar 177/10
🧠

CRASH: Cognitive Reasoning Agent for Safety Hazards in Autonomous Driving

Researchers introduced CRASH, an LLM-based agent that analyzes autonomous vehicle incidents from NHTSA data covering 2,168 cases and 80+ million miles driven between 2021-2025. The system achieved 86% accuracy in fault attribution and found that 64% of incidents stem from perception or planning failures, with rear-end collisions comprising 50% of all reported incidents.

AIBullisharXiv – CS AI · Feb 277/107
🧠

OmniGAIA: Towards Native Omni-Modal AI Agents

Researchers introduce OmniGAIA, a comprehensive benchmark for evaluating omni-modal AI agents that can process video, audio, and image data simultaneously with complex reasoning capabilities. They also propose OmniAtlas, a foundation agent that enhances existing open-source models' ability to use tools across multiple modalities, marking progress toward more capable AI assistants.

AINeutralarXiv – CS AI · Mar 266/10
🧠

GameplayQA: A Benchmarking Framework for Decision-Dense POV-Synced Multi-Video Understanding of 3D Virtual Agents

Researchers introduce GameplayQA, a new benchmarking framework for evaluating multimodal large language models on 3D virtual agent perception and reasoning tasks. The framework uses densely annotated multiplayer gameplay videos with 2.4K diagnostic QA pairs, revealing substantial performance gaps between current frontier models and human-level understanding.

AIBullisharXiv – CS AI · Mar 37/108
🧠

Nano-EmoX: Unifying Multimodal Emotional Intelligence from Perception to Empathy

Researchers have developed Nano-EmoX, a compact 2.2B parameter multimodal language model that unifies emotional intelligence tasks across perception, understanding, and interaction levels. The model achieves state-of-the-art performance on six core affective tasks using a novel curriculum-based training framework called P2E (Perception-to-Empathy).

AIBullisharXiv – CS AI · Mar 36/107
🧠

Dr. Seg: Revisiting GRPO Training for Visual Large Language Models through Perception-Oriented Design

Researchers introduce Dr. Seg, a new framework that improves Group Relative Policy Optimization (GRPO) training for Visual Large Language Models by addressing key differences between language reasoning and visual perception tasks. The framework includes a Look-to-Confirm mechanism and Distribution-Ranked Reward module that enhance performance in complex visual scenarios without requiring architectural changes.

AINeutralarXiv – CS AI · Mar 24/105
🧠

TaCarla: A comprehensive benchmarking dataset for end-to-end autonomous driving

Researchers have released TaCarla, a comprehensive dataset containing over 2.85 million frames from CARLA simulation environment designed for end-to-end autonomous driving research. The dataset addresses limitations in existing autonomous driving datasets by providing both perception and planning data with diverse behavioral scenarios for comprehensive model training and evaluation.

$RNDR