y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#cognitive-abilities News & Analysis

6 articles tagged with #cognitive-abilities. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

6 articles
AINeutralarXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

The ARC of Progress towards AGI: A Living Survey of Abstraction and Reasoning

A comprehensive survey of 82 AI approaches to the ARC-AGI benchmark reveals consistent 2-3x performance drops across all paradigms when moving from version 1 to 2, with human-level reasoning still far from reach. While costs have fallen dramatically (390x in one year), AI systems struggle with compositional generalization, achieving only 13% on ARC-AGI-3 compared to near-perfect human performance.

๐Ÿง  GPT-5๐Ÿง  Opus
AINeutralarXiv โ€“ CS AI ยท Mar 47/104
๐Ÿง 

A Neuropsychologically Grounded Evaluation of LLM Cognitive Abilities

Researchers introduced NeuroCognition, a new benchmark for evaluating LLMs based on neuropsychological tests, revealing that while models show unified capability across tasks, they struggle with foundational cognitive abilities. The study found LLMs perform well on text but degrade with images and complexity, suggesting current models lack core adaptive cognition compared to human intelligence.

AIBearisharXiv โ€“ CS AI ยท Mar 266/10
๐Ÿง 

Visuospatial Perspective Taking in Multimodal Language Models

Research reveals that multimodal language models have significant deficits in visuospatial perspective-taking, particularly in Level 2 VPT which requires adopting another person's viewpoint. The study used two human psychology tasks to evaluate MLMs' ability to understand and reason from alternative spatial perspectives.

AINeutralarXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

To Think or Not To Think, That is The Question for Large Reasoning Models in Theory of Mind Tasks

A research study of nine advanced Large Language Models reveals that Large Reasoning Models (LRMs) do not consistently outperform non-reasoning models on Theory of Mind tasks, which assess social cognition abilities. The study found that longer reasoning often hurts performance and models rely on shortcuts rather than genuine deduction, suggesting formal reasoning advances don't transfer to social reasoning tasks.

AINeutralarXiv โ€“ CS AI ยท Apr 75/10
๐Ÿง 

When Models Know More Than They Say: Probing Analogical Reasoning in LLMs

Researchers found that large language models (LLMs) have an asymmetry between their internal knowledge and prompted responses when detecting analogies. While probing reveals models understand rhetorical analogies better than their prompted responses suggest, both methods perform poorly on narrative analogies requiring deeper abstraction.